Tuesday, February 23, 2021

NCAA TOURNAMENT: KEEPING THE RPI HONEST - PART 1

There are two big questions for the Women’s Soccer Committee this Spring:

1.  Will we be able to use the RPI as a tool when forming the NCAA Tournament bracket?  If so, how useful will it be?

2.   If we can’t use the RPI or if its usefulness is limited, how will we form the bracket?

In my November 27, 2020 article, I used the RPI rankings during the Fall season to show how the RPI works.  For the RPI this year, there are three areas of concern: the reduced number of games teams will be playing, the reduced proportion of non-conference games, and the reduced proportion of out-of-region games.  And in my December 26, 2020 article, I discussed potential non-RPI based tools for forming the bracket.  These two topics will be a main focus for me over the coming weeks.

 According to the Pre-Championship Manual, the NCAA will do RPI calculations for the 2020-21 season.  It will issue its first RPI reports the week of March 1 and will issue reports weekly thereafter.  The purpose of these reports this year mainly may be to produce the Team Sheets that the Committee uses as a tool during bracket formation.  The Team Sheets contain detailed information about each team’s record, above and beyond RPI-related information.  The planned issuance of RPI reports suggests, however, that the NCAA and the Committee may not have decided yet whether the RPI ratings and ranks themselves will be a usable bracket formation tool.

To help see whether the RPI will be usable, I created two tests to apply to the RPI rankings as the NCAA issues them weekly.  I will explain the tests in this post (Part 1).  In the next post (Part 2), I will will apply the tests to simulated end-of-season rankings I have done.  You will be able to use your own judgment as to whether it is likely the RPI will be usable.

For a basic data resource, the tests use the end-of-season RPI rankings for the years 2013 to the present.  (2013 is the year of completion of the most recent major conference membership re-alignment.)  For each year, the first test’s data are the rankings and conferences of the Top 60 teams.  The second test’s data are the same but limited to the Top 30 teams. I use the Top 60 because all Tournament at large teams since 2007 have come from the Top 57, and the Top 60 is a nice rounded off number.  I use the Top 30 because since 2007, all RPI Top 30 teams have gotten at large selections.

Top 60 Test

Below are three tables.  I will explain below each one.


This table covers all of the conferences that have had at least one team in the Top 60 in at least one of the years.  In determining the Top 60, I have excluded Ivy League and Big West teams, since they are not playing in 2020-21, and I have replaced them with the next teams in the rankings to bring the total teams for each year to 60. The green highlighted conferences are the ones that have had teams in the Top 60 every year.  At the bottom of the table, the last two rows show the number of teams each of the highlighted and not highlighted conferences had, for each year.  You can see that the numbers for the highlighted and not-highlighted conferences are quite consistent from year to year.


This table shows the average yearly number of teams the conferences have had in the Top 60.  Again, the green highlights are for the conferences that have had teams in the Top 60 every year.


This table shows the averages for the highlighted and not-highlighted groups.  Rounding off, of the RPI Top 60 teams, 49 have been from the highlighted group (including their automatic qualifiers) and 11 from the not-highlighted group (6 automatic qualifiers plus 5 more).

Using these numbers as a test for the reasonability of the RPI this year, we reasonably can expect that the RPI Top 60 will include roughly 49 teams from the highlighted conferences and 11 teams from the not-highlighted conferences.  Further, using the minimum and maximum numbers from the bottom of the first table, we can expect that 45 teams will be the minimum we should see from the highlighted group and 15 the maximum from the not-highlighted group.

Top 30 Test

The following three tables are the same as the above three, except are based on the Top 30 in the RPI rankings.





Summarizing these three tables: Using the numbers as a test for the reasonability of the RPI this year, we reasonably can expect that the RPI Top 30 will include roughly 28 teams from the highlighted group and 2 teams from the not-highlighted group.  Further, using the minimum and maximum numbers from the bottom of the first table, we can expect that 27 teams will be the minimum we should see from the highlighted group and 3 the maximum from the not-highlighted group.

Tests Summary

The following tests should be applied to each NCAA RPI report:

Top 60 Test:  The RPI Top 60 should include roughly 49 teams from the highlighted conferences and 11 teams from the not-highlighted conferences.  The actual numbers can range on either side of these test numbers, but 45 teams should be the minimum from the highlighted group and 15 the maximum from the not-highlighted group.

Top 30 Test:  The RPI Top 30 should include roughly 28 teams from the highlighted conferences and 2 teams from the not-highlighted conferences.  The actual numbers can range on either side of these test numbers, but 27 teams should be the minimum from the highlighted group and 3 the maximum from the not-highlighted group.

What to Watch For

I expect that the early RPI rankings will not match well with these "reasonable expectation" test standards.  I also expect the weekly rankings that follow will do better in relation to the test standards.  If I am right, the ultimate question will be whether the final RPI rankings have gotten close enough to the test standards to indicate they have at least some usefulness and, if so, how much.

I will be watching in particular for three things:

1.  Are too many teams from the not-highlighted conferences showing up in the Top 30 and Top 60?

2.  Are the distributions of the teams among the highlighted conferences reasonably consistent with the averages those conferences have had since 2013?

No comments:

Post a Comment