CIG Elo ratings
Elo as calculated by Remi Coulom’s bayeselo
program. The # column gives the official ranking, so you can see how it differs from the rank by Elo. The bayeselo
ranking should be slightly more accurate because it takes into account all the information in the tournament results, but unfortunately there are missing games so the Elo is computed from slightly less data than the official results. The “better?” column tells how likely each bot is to be superior to the one below it.
# | bot | score | Elo | better? |
---|---|---|---|---|
1 | ZZZKBot | 75% | 1749 | 98.8% |
2 | tscmoo | 74% | 1731 | 100% |
3 | PurpleWave | 67% | 1660 | 99.9% |
4 | LetaBot | 63% | 1626 | 97.9% |
5 | UAlbertaBot | 62% | 1611 | 85.1% |
6 | MegaBot | 61% | 1604 | 96.7% |
7 | Overkill | 60% | 1591 | 89.3% |
8 | CasiaBot | 59% | 1582 | 55.2% |
9 | Ziabot | 59% | 1581 | 58.8% |
10 | Iron | 58% | 1579 | 97.0% |
11 | AIUR | 57% | 1566 | 100% |
12 | McRave | 47% | 1476 | 97.6% |
13 | Tyr | 45% | 1462 | 79.9% |
14 | SRbotOne | 45% | 1456 | 99.9% |
15 | TerranUAB | 38% | 1397 | 99.9% |
16 | Bonjwa | 33% | 1347 | 94.9% |
18 | OpprimoBot | 32% | 1335 | 69.0% |
17 | Bigeyes | 32% | 1331 | 99.9% |
19 | Sling | 26% | 1275 | 100% |
20 | Salsa | 9% | 1041 | - |
Looking at the better? column, we see that the top 3 are well separated; the places are virtually sure to be accurate. ZZZKBot and Tscmoo are close, but bayeselo
thinks they are separated enough. Farther down, CasiaBot, Ziabot, and Iron are statistically hard to distinguish; there is not strong evidence that they finished in the correct order. Also OpprimoBot and Bigeyes are not well separated—as you might guess since their results are reversed from the official results.
Is this all the analysis we want of CIG 2017? I also have a script for the map balance, to check whether any race is favored. But it tells more about who competed than about the maps or bot skills.
Comments