Top 25  Truncates every computer rating's ranking at 25 and then counts the
ballots the same way media polls do, using a 252423... point assignment for teams ranked 123....
I only include this report to demonstrate how much information is left out of the usual media presentation of their poll results.
In addition to the number of "points" I include the number of ballots that listed the team in the top 25 and the number of votes
for each rank for which the team was voted in the top 25.

Top 25 Correlations  See below for a general discussion of rankcorrelations. The top 25 correlations are a bit different, since not every ranking includes the same teams. The reports are described in Top 25 Correlations.

Borda  The usual method of counting top 25 ballots is a variation on
the Borda Count. In its basic form, teams get one point for each team they are ranked better than. In a 130team field,
a #1 vote is worth 129, #2 worth 128 down to #130 worth 0. When all teams are ranked the order is the same as the average
rank over all ratings.

Majority Consensus 
My consensus rank is based upon the Bucklin votecounting method. For each team find the best rank for which a majority
of the ratings agree the team should be ranked at least that highly. I use a strict majority, namely 50% + one rating.
When there are an odd number of ratings this is the same as the arithmetic median. For an even number of ratings it is
the best rank worse than the median.
Unless otherwise stated this is the ranking used in reports that include a team's rank.

Pairwise Matrix  Even when the majority ranks team A better than
team B, it is possible that team B is ranked better on more ballots than team A. In Condorcet voting, the ballots are
translated into pairwise comparisons between alternatives. The method suffers from a lack of transitivity: team A > team B and
team B > team C does not imply team A > team C!

Playoff Approval 
Approval voting does not require a ranked list. Instead the voter just lists the alternatives that meet or exceed their minimum criteria. For division 1A football the criterion for selecting the four playoff teams is "one of the best four teams in the field." As I suggested in Committee "consensus" this would be a much better way for the committee to amalgamate its collective thoughts than the "top25" poll they chose. My report just counts the number of voters that rank the team in its top 4.

Correlations 
The basis for measuring how alike two ordinal rankings
are is the distance metric. This is the number of swaps required by a bubble sort to place one of the lists in the
same order as the other. The distance varies from zero (the lists are identical) to the total number of teampairs (the lists
are reversals of each other; 8,128 for a 128team field.) For each ranking I report the contribution to the distance function by each team.
The distance is the number of discordant pairs  the number of pairs where the teams' relative orders are reversed in the
two rankings. When the teams are in the same relative order in both lists the pair is said to be concordant. When
the teams have the same rank in either list the pair is ignored.
These can be turned into rank correlation coefficients in several ways. The two I calculate are:
Kendall's tau:
τ =  #Concordant pairs  #Discordant pairs 

½ × #Ranked × (#Ranked1) 
Goodman and Kruskal's gamma:
γ =  #Concordant pairs  #Discordant pairs 

#Concordant pairs + #Discordant pairs 
These give 1 ≤ { τ, γ } ≤ 1 with τ ≤ γ. Both will be 1 if the teams are in exactly reverse order, 0 if the relationship
is perfectly random (whatever that means!) and +1 if the rankings are identical. The τ and γ are the same
if there are no ties (but notice that ties in the Majority Consensus rank are to be expected, in which case τ will be closer to zero
than γ.)

Conference Ranks 
There are more ways to aggregate team ranks by conference than ratings, but I have chosen these.
 Rank Distribution by Conference
 includes every rank for every team in the conference. In addition to the average
rank there's a count of team ranks in the ranges, 125, 2650, and so on.
 Team Consensus Ranks by Conference
 lists the consensus ranks of teams by conference.
 Pairwise Comparison of Teamranks by Conference
 compares the consensus rank of each conference team to that of every member of other conferences. The
entries represent the number of times out of a thousand that a team from the row conference would be expected to
have a better rank than that of a team from the column conference.

Weighted Violations  Roughly one in five games result in the
worseranked team winning no matter which rating produces the ranking. Were it not so sport would not be interesting. One
measure of how well a rating represents resultstodate is the count of Retrodictive Ranking Violations, the number
of games in which after the ranking takes into account team A beat team B it still ranks team B better than team A.
Motivated by Potemkin's idea that instead of just counting the
number of RRVs we shuold take into account the size of the violation (rank difference) and the importance (how
highly the loser is ranked) I came up with a Weighted RRV value that combines the size of the upset (in scores and
rank difference) with importance (loser's rank.) The "importance" component also takes into account that violations later in the
season (when the rating has more input) should count more.
WRRV = ⌈ (WS  LS)÷S ⌉ × (WR  LR) × ƒ(LR) 
 WS
 Winner's score
 LS
 Loser's score
 ⌈ (WS  LS)÷S ⌉ =

the margin of victory in number of scores
(⌈x⌉ is the least integer ≥ x)
S=8 for football
 WR
 Winner's rank
 LR
 Loser's rank
Note that (WRLR) is positive for all ranking violations by definition.
 ƒ(LR) =

Log _{ LRi Mj} M^{i+j}
where i and j are chosen to make ƒ(1) any predetermined value ƒ(M) is exactly 1. M is the number of ranked teams (130 in 2017 for Division 1A.)

See WRRV 2.1 for graphs of ƒ(LR) for different choices of i and j.
