My definition of the "consensus" team rank is the best rank for the team for which a majority of rating systems rank the team at least that highly. That is the median rank if the number of ratings is odd, or the best rank worse than the median for an even number of ratings. (You can find it by alternately eliminating the best and worst of remaining ranks, beginning with the best, until there is only one ranking left.)
The "correlation to consensus" I use is the "distance" of a ranking from the consensus ranking. This is the number of pairs in the opposite order than in the Majority ranking. I like this measurement of correlation because it has a simple interpretation: it is the number of swaps a bubble sort would require to transform the ranking into the consensus ranking.
The distance is the number of discordant pairs - the number of pairs where the teams' relative orders are reversed in the two rankings. When the teams are in the same relative order in both lists the pair is said to be concordant. When the teams have the same rank in either list the pair is ignored.
These can be turned into rank correlation coefficients in several ways. The two I calculate are:
Kendall's tau:
Goodman and Kruskal's gamma:
τ = #Concordant pairs - #Discordant pairs # Total Pairs
These give -1 ≤ τ ≤ γ ≤ 1. Both will be -1 if the teams are in exactly reverse order, 0 if the relationship is perfectly random (whatever that means!) and +1 if the rankings are identical. The τ and γ are the same if there are no ties (but notice that ties in the Majority consensus rank are to be expected.)
γ = #Concordant pairs - #Discordant pairs #Concordant pairs + #Discordant pairs
|
Another feature of the distance metric that makes it highly useful is that it is possible to capture and report the contribution of individual team ranks by a rating to the size of the variation. I've added a Corr report to the Analysis: links on the home page to report those, with ratings listed in closest-to-farthest from consensus order.