The usual image of an artificial neural network:
What follows is a paper that I wrote in the Spring of 2001 for an “Introduction to Neural Networks” class that I took as part of my Master’s degree. It is mostly a review of someone else’s paper on the subject, except that I wrote the network in Excel and ran it on that year’s football season games. Fun, fun.
Abstract
This paper describes an approach to the problem of ranking college football teams that determines two competitors for the National Championship. A competitive NN was built, validated, and trained with a bivalent purpose: to rank the teams according to their performance during the regular season and, based on that raking, to predict results. A comparison with other computer rankings will show the good approximation achieved by a NN ranking that predicted 80 % of the results at the end of the season.
Introduction
The National College Football Championship is defined every January by the two best teams from Division I-A, which is the most important. The regular season is played in 15 consecutive weeks between August and December and 115 teams compete for a place in the final. Since in that very short time teams play at most 12 games, it’s impossible to get a complete all-round. Therefore every team has its own schedule very different from other teams and the wins-losses record is not comparable (even if two teams have, say, a 9-2 record, that is achieved playing different opponents).
The method used so far to rank them, shown below, uses the media of two subjective rankings by Associated Press (poll among coaches) and USA Today (poll among sportswriters), the difficulty of the schedule/25, number of lost games and the average position in some reputed computer rankings.
Poll | Sched | Comp | ||||||
Team | AP | USA | AVG | SS | RANK | L | AVG | Total |
Oklahoma | 1 | 1 | 1 | 11 | 0.44 | 0 | 1.86 | 3.3 |
Florida St | 3 | 3 | 3 | 2 | 0.08 | 1 | 1.29 | 5.37 |
Miami Fl | 2 | 2 | 2 | 3 | 0.12 | 1 | 2.57 | 5.69 |
Washington | 4 | 4 | 4 | 6 | 0.24 | 1 | 5.43 | 10.67 |
Virginia Tech | 6 | 5 | 5.5 | 14 | 0.56 | 1 | 5.14 | 12.2 |
Highlighted on the right are the final scores for Florida State and Miami Florida. The narrowness of the difference brings controversy and complaints from Miami Florida, relegated from playing the final, especially because during the season, they defeated Florida State. This causes lobbying and politics, which affect the nature of the game every year. A method objective and reliable would end these problems. As an example, the calculation for Oklahoma is 1+0.44+0+1.86=3.3. The table also shows Oklahoma and Florida State as top-ranked and chosen to play the final, the Orange Bowl, next January 3rd.
Network Architecture
A competitive network suits perfectly for this job. Competitive networks provide a winner but in this case, although matches give only one winner, the network is arranged so that the “winner won’t take all”. Shall we do that then the net would only tell me who is number one, but nothing about the rest.
The net consists of 115 nodes. Each node represents a team. The value of the node is the value we’ll use to rank the teams.
Nodes connected are the nodes representing teams who played each other during the season: if two teams haven’t played, then the two nodes representing them won’t be connected. Since every team plays about 10 games per season, then each node will be connected to the other 10 nodes among the 115. This gives a low-populated network.
The weight of the connection has a bivalent value explained by the figure:
In this net, “B” defeated “A” by 12 points: the connection with “B” is giving “A” a value of –12 for that game, and the connection with “A” is giving “B” a value of +12 for that game. Also “C” and “A” played each other (B won), whereas “A” and “C” didn’t play so they’re not connected.
In order for the ranking to be valid, the net has to be fully connected (at least, a minimum spanning tree). Meaning is that if all the teams are not connected, then the result is more than one ranking, each including only teams connected together.
The net could have been connected in only two weeks, with each team playing only two games (also net could be not fully connected even after up to 57 weeks, though in this season teams became connected after 7 weeks).
Once the net is validated (by being fully connected, an objective condition) has to be reliable. The more games played per team, the more reliable the net becomes. In this season, teams played between 6 and 12 games, with the average between 10 and 11 (this is a subjective condition: the average of 10 games could be enough but more is better).
Data Collection
Data are the results of the games between Division I-A teams during the first 12 weeks of the regular season 2000/01. All data comes from sports sites on the web and all results available were used (no sampling). Data is highly reliable after several checks performed in a spreadsheet.
Since the data input in the net is very time-consuming and the season has just ended, the net was constructed using results up to the 12th week (played past November 11th). In those 12 weeks, the 115 teams played 541 games. In this way, 1082 weights were input. As said before: since the net has to know who won the match and by what margin, one positive weight is allocated to the winner and one negative weight to the loser.
Transfer Function
The transfer function has the form:
VG=A+B * (if W>0,1,0) + [a+b * (if W>0,1,0)]*VO
The value of a node is the average transferred from connections (the value of a team is the average of the value it got from all games played).
In an iterative process, the transfer function is run until convergence is assured. The stopping rule was very simple: if two consecutive rankings are the same, then stop. In this experiment, the process proved to be highly convergent and 20 iterations were always enough.
Training
Training parameters of the transfer function (and keeping weights fixed) A, B, ( and ( we pursue maximizing the percentage of results predicted. For this experiment, the net is built with 541 results from weeks 1-12, trained using 103 results from weeks 11 and 12, and tested over 73 games on weeks 13 to 15.
Although net could have been trained using all 541 games from weeks 1-12 only the last two weeks were used because:
-that would have represented a big amount of work.
-training with the last two weeks let us give more importance to the last games played for a team: the last games give a better idea of how the team is going to perform next time. Finlayson uses what he calls the “time weighting factor” that improves his predictions for international rugby games.
A Home Advantage Factor (which would have given the results as if played in the neutral field) could not be implemented because it permanently provoked fluctuations in the iteration trend.
To assure convergence, parameters were chosen so that the value of a game was always between 0 and 100.
Results
Using parameters: A = 30; B = 20; = 0.35; = 0.5 and giving an initial value of 50 for the iteration (bias) to all, 80% of results are correctly predicted. The complete ranking is shown as an appendix.
As a reference, Dr. Ray Finlayson’s world rugby ranking predicts 77% of results and 84% after the time weighting.
Unfortunately, a comparison with other college football rankings is not possible since they usually only provide the top 16 or 25 teams, but all of them are needed.
Neural Network Football Ranking
The following table depicts the correlation between the most reputed computer rankings and the one proposed here. Teams are sorted by the BCS (official) ranking. Highlighted cells show how the “bugs” incurred by the ANN ranking have correspondences in the other rankings: Washington ranked 23rd by the ANN and 4th by the BCS, is ranked 10th by Billingsley and 11th by Dunkel; Nebraska, ranked 4th by the ANN and 8th by the BCS, was also 4th according to Sagarin, etc.
BCS Ranking | BIL | DUN | MAS | NYT | ROT | SAG | SCR | SEA | ANN |
Oklahoma | 1 | 3 | 2 | 3 | 1 | 3 | 2 | 1 | 2 |
Florida St | 2 | 1 | 1 | 1 | 2 | 1 | 1 | 3 | 1 |
Miami Fl | 3 | 2 | 3 | 2 | 3 | 2 | 3 | 4 | 3 |
Washington | 10 | 11 | 5 | 5 | 4 | 8 | 4 | 2 | 23 |
Virginia Tech | 5 | 5 | 4 | 4 | 7 | 5 | 7 | 6 | 6 |
Oregon St | 7 | 9 | 8 | 8.5 | 5 | 7 | 5 | 5 | 11 |
Florida St | 4 | 4 | 7 | 6 | 9 | 6 | 6 | 7 | 5 |
Nebraska | 6 | 13 | 6 | 10 | 6 | 4 | 8 | 9 | 4 |
Kansas St | 8 | 12 | 11 | 12 | 8 | 9 | 11 | 12 | 8 |
Oregon | 12 | 17 | 14 | 15 | 11 | 14 | 9 | 8 | 19 |
Where the headers mean: Billingsley, Dunkel, Massey, New York Times, Rothmans, Sagarin, Scripps, Seattle Times, the ANN discussed here
Future Work
- Develop a net in which the difference in the ranking gives the expected score difference in a game between that teams (as Finlayson and Sagarin do)
- Rewrite the Home advantage factor so that it does not provoke fluctuations in the trend (Sagarin).
- Improve the time weight factor, to give more value to a game played last week as compared to any game played before (Finlayson).
- Develop a parameter to give more value to games played between top-ranked teams.
Conclusion
The objectives targeted by this work were partially accomplished. The 80% prediction sounds exciting and overall the rank shows being very similar to other computer rankings.
A deficit is a failure in implementing the H/V factor, which gave instability to the convergence of the iterations.
Nomenclature
VG = value transferred by the connection (the value the team gets from the game)
VO = value of the node connected (value of the opponent team) (since this is an iterative process VO is VG for the opponent in the previous iteration)
W = weight of the connection (score difference of the game)
A + B = intercept for the “win” curve
A = intercept for the “loss” curve
a + b = slope for the “win” curve
a = slope for the “loss” curve
References
Fausett, L. (1994). Fundamentals of Neural Networks. First Edition. Upper Saddle River. Simon & Schuster.
Wilson, R. (1995). “Ranking College Football Teams: A Neural Network Approach”. Interfaces 25: 4 July-August 1995 (pp. 44-59).
On the Internet:
http://www-cs-students.stanford.edu/~rsf/finlayson.html; http://sports.yahoo.com; http://www.ncaa.org; http://collegefootball.net; http://www.planet-rugby.com/rankings/explained.html/PR/ENG; http://www.live.com/wrr/
Other College Football Rankings:
http://www.usatoday.com/sports/sagarin.htm; http://www.dunkelindex.com; http://www.mratings.com/rate/cf-m.htm; http://www.cae.wisc.edu/~dwilson/rsfc/rate/rothman.txt; http://espn.go.com/abcsports/bcs; http://college.espn.go.com/ncf/rankings
Appendix
Final ANN Ranking:
1 | Florida St. | 22.04 |
2 | Oklahoma | 30.18 |
3 | Miami Fla. | 32.06 |
4 | Nebraska | 32.54 |
5 | Florida | 34.61 |
6 | Virginia Tech | 36.51 |
7 | Toledo | 36.57 |
8 | Kansas St. | 36.91 |
9 | Texas Christian | 37.39 |
10 | Purdue | 38.38 |
(actually, 115 teams were ranked)
**************
That was the paper, over time I decided the network actually more closely resembles a Markov Chain than a typical neural network (feedforward et al) where the “states” are the teams and the probabilities are games’ scores.
Years later, in 2007, I joined the crowd of computer rankings enthusiasts and created SporTheory, my very own page where I publish the rankings resulting from these techniques and rank college teams in Football and Basketball. I used to also rank Baseball, Softball and Hockey teams.
If you are curious about neural networks, check the textbook we used:
Thanks for the rundown. This was super interesting. I think it would be fun if you did an update analysis on how the College Football Playoffs changed any trends you were seeing before with schedule, more games.
Cheers
I got rid of sportheory.com and now the ranking is on this site: https://kindoflost.com/college-football-ranking/
You can also find other rankings, created with the same approach, from the top menu
Liked the post! Love to see if you think the College Football Playoffs have impacted any trends you were seeing.
Cheers