Results vs Opponents' Ratings

August 30, 2015

In Ratings Scope I introduced my new ISR and ISOV report format, which compared to prior years adds the "Norm" column and eliminates the "ASOS" and "PASOS" meta-ratings. I mentioned that I'd found histograms to be a better tool for this kind of analysis.

We're familiar with the tables that show "records vs top 25, 50, …" but I do not find those very useful because the ranks are arbitrary and the differences between the extremes of the ranges is so variable. There's a much larger difference between #25 and #1 than there is between #50 and #26.

Instead for "buckets" I divide the field into ½ standard deviation (σ) -wide groups of teams then apply a basic "bonus for better wins, penalty for worse losses" formula. This is "vs opponents' ratings" as opposed to opponents' ranks.

When applied to the 2014 season (with the field modified as described in the cited article) for the ISR this results in a "records vs teams ranked 1-4, 5-17, 18-35, 36-130, 131-162, 163-180, 181-196, 197+". The widest rank range corresponds to opponents within ½ standard deviation of average, and both wins over and losses to teams in this range are assigned value 1.

Win-Loss Weight Example for 2014 ISR

low*σhi*σAdd for winSub for lossOpp ISR>Low RankHigh Rank#teams
2+∞40.0625129.29144
1.5230.125121.5851713
11.520.25113.86183518
0.511.50.5106.14366429
-0.50.51190.716513066
-1-0.50.51.582.9913116232
-1.5-10.25275.2716318018
-2-1.50.125367.5518119616
-∞-20.06254 -∞1971982

The weightings are arbitrary, but the only requirements for them are that they be symetrical around the center bucket and be significantly different enough at the tails to distinguish between "good/bad" and "even better/worse" in the formula's result. Again using 2014 results with the modified field, for the ISR the report looks like this.

Records vs Opponents' ISR Values

30 Aug 2015 11:39am (US Mountain)

SortRankISRTeamRecConf>2σ>3σ/2>σ/2>-σ/2>-σ>-3σ/2>-2σ>-∞
26.750011Ohio State 14-1B10 2-0 2-0 1-0 6-1 2-0 0-0 1-0 0-0 0-0
24.312522Oregon 13-2P12 1-1 3-1 2-0 2-0 4-0 1-0 0-0 0-0 0-0
19.687533Florida State 13-1ACC 0-1 1-0 4-0 3-0 4-0 0-0 1-0 0-0 0-0
19.312544Alabama 12-2SEC 0-1 1-1 6-0 2-0 0-0 3-0 0-0 0-0 0-0
18.437555UCLA 10-3P12 0-1 3-0 1-2 4-0 2-0 0-0 0-0 0-0 0-0
17.625066TCU 12-1B12 0-0 1-1 3-0 3-0 4-0 0-0 1-0 0-0 0-0
16.312578Missouri 11-3SEC 0-1 0-1 4-0 5-0 2-1 0-0 0-0 0-0 0-0
16.125089Mississippi 9-4SEC 1-0 1-1 2-3 2-0 3-0 0-0 0-0 0-0 0-0
16.0625914Arizona 10-4P12 1-1 1-3 1-0 2-0 4-0 1-0 0-0 0-0 0-0
...
-11.5000192187UC Davis 1-9BSky 0-0 0-0 0-2 0-0 1-3 0-2 0-0 0-2 0-0
-11.5000192178New Mexico State 2-10SBC 0-0 0-0 0-1 0-0 1-6 0-1 1-1 0-1 0-0
-12.5000194183Weber State 2-10BSky 0-0 0-1 0-1 0-0 0-3 0-3 1-1 1-1 0-0
-12.6875195188Norfolk State 4-8MEAC 0-0 0-0 0-0 0-0 0-2 1-5 0-0 2-0 1-1
-13.4375196195Hampton 2-8MEAC 0-0 0-0 0-0 0-0 0-2 1-4 0-0 0-2 1-0
-16.3125197197Delaware State 2-8MEAC 0-0 0-0 0-0 0-0 0-1 0-3 0-1 1-3 1-0
-20.5000198198Savannah State 0-10MEAC 0-0 0-0 0-0 0-0 0-3 0-3 0-0 0-3 0-1
6-52 37-137 73-161 144-227 406-382 189-119 152-45 127-27 18-2

SortThe value of the formula applied to the teams' record is not by itself very useful, but is included to make it somewhat easier to analyze a pair of schedules/results.
Rank
rating
The relative rank of each team according to this better-win/worse-loss formula. Teams with the same sort value will have rank one greater than the team(s) with the next higher sort value.
rating (ISR in the example) is the team's relative rank according to the base rating.

The degree to which the two rankings agree is a measure of how retrodictively self-consistent the rating is. When a pair of teams' relative position is opposite in the two rankings, it should be easy to recognize an "upset" in one of the team's rows in the table.

TeamThe team's name. By the time I publish my first rankings this will very likely be a link to a more detailed analysis of the team's results.
RecTotal wins and losses against teams in the field. Games that were not used to calculate rating are treated as if they did not occur.
ConfThe team's conference affiliation. Mainly included to provide a visual break between the overall record and the histogram table.
>2σ
   to
>-∞
Team's record vs teams whose rating values are better than the average plus the number of standard deviations indicated by the column headings.

>3σ/2 should be read as "opponents' rating greater than μ+1.5×σ but less than μ+2×σ", where μ is the rating average and σ its (population) standard deviation.

Application to D1A Team Rankings
In cases where we do not have access to a rating's values but have the ranking by that rating (e.g. from Dr. Massey's rankings list) we can assume that the rating values are normally distributed (they are for all "advanced" systems, and the others we don't care about.) A quick look at the probability density function for the normal distribution suggests how to form the buckets for the histograms:

SD range%in range×128Lo RankHi Rank
> 22.2752.9113
1.5 to 24.4065.6449
1 to 1.59.18511.761021
0.5 to 114.98819.182240
-0.5 to 0.538.29249.014189
-1 to -0.514.98819.1890108
-1.5 to -19.18511.76109120
-2 to -1.54.4065.64121126
< -22.2752.91127128

So instead of records vs 1-25, 26-50, etc. for the 128-team 1A field there is more value derived by using 1-3, 4-9, 10-21, 22-40, 41-89, 90-108, 109-120, 121-126 and 127+. In fact if we had a perfect computer rating (call it Greatest Of Deciders) and still wanted human involvement in selecting the 4-team tournament teams it would make a lot of sense to let the G.O.D. pick the first three teams and only ask the humans to pick which of the other six teams ranked better than 10th should complete the tournament.

© Copyright 2015, Paul Kislanko