The method I use to consolidate computer rankings to form a "meta-ranking" is similar to a voting system known as "Bucklin". It is less subject to extreme rankings than a simple arithmetic mean or the method typically used in sports polls (Borda counts).
The basic idea is simple - assign a team rank R if a majority of voters (or computer rankings, in this case) agree that the team should have rank R or a better rank. Votes (or computer rankings) for a better rank are counted for the rank where a majority agrees the team should be ranked at least that high.
This vote-counting method is not affected by extreme rankings in either direction. If someone (or some computer) ranks Ivy-Covered-U first, but all other voters rank ICU 50th, the vote for #1 gets counted as a vote for #50. Likewiise if some mean-spirited voter (or biased compter) ranks ICU very low, that becomes irrelevant because votes below the "majority agrees" rank don't affect ICU.
There can be (and usually are) ties if all we consider is the highest ranking for which a majority agrees is appropriate.
Maj | Cnt of 16 | Best | Worst | Team | Conf | HOW | KEE | CLA | MAS | DWI | MOR | BIL | DES | DOK | GM | UCS | PFZ | CGV | CSL | SOL | THM | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7 | 10 | 10 | 4 | 14 | Georgia | SEC | 6 | 5 | 11 | 10 | 13 | 12 | 6 | 13 | 10 | 10 | 10 | 14 | 9 | 10 | 12 | 4 |
8 | 10 | 9 | 5 | 32 | Boise St | WAC | 10 | 15 | 8 | 17 | 8 | 20 | 9 | 20 | 5 | 31 | 32 | 12 | 6 | 6 | 10 | 10 |
Maj | Cnt of 16 | Best | Worst | Team | Conf | HOW | KEE | CLA | MAS | DWI | MOR | BIL | DES | DOK | GM | UCS | PFZ | CGV | CSL | SOL | THM | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14 | 14 | 10 | 2 | 18 | Tennessee | SEC | 15 | 9 | 14 | 16 | 18 | 14 | 14 | 2 | 7 | 6 | 15 | 18 | 15 | 12 | 4 | 7 |
15 | 14 | 10 | 4 | 29 | Ohio State | B10 | 14 | 11 | 22 | 14 | 29 | 10 | 16 | 9 | 16 | 4 | 7 | 23 | 5 | 19 | 6 | 5 |
One of the byproducts of this measurement is that we can rank the rankings based upon the percentage of times the rankings are a part of the majority that decides a team ranking. This is at least interesting;
DOK | 80 | 0.672 |
BIL | 74 | 0.622 |
DWI | 73 | 0.613 |
CGV | 72 | 0.605 |
HOW | 71 | 0.597 |
PFZ | 71 | 0.597 |
MAS | 70 | 0.588 |
GM | 69 | 0.580 |
THM | 69 | 0.580 |
KEE | 68 | 0.571 |
DES | 68 | 0.571 |
SOL | 68 | 0.571 |
MOR | 66 | 0.555 |
CSL | 65 | 0.546 |
CLA | 65 | 0.546 |
UCS | 64 | 0.538 |
This is not a very strong correlation, because it is "one-sided." A rating that assigns a rank that is "too high" is not counted as an error for the same reason it does not contribute to the composite being "too high". Nonetheless, at the top of the rankings there are fewer choices for erroneously high ranks, so where it matters most this simple counting statistic suffices.