The formula that reveals this season's AFL ladder

By Nick Croker / Roar Guru

Prior to the season I made a little algorithm to try and predict the ladder.

I didn’t manage to tap out an article in time and many other people put their own predictions up, which led me to think such an article might be redundant.

However, I like it as a tool.

The results over a long data sample are fairly impressive.

Having said that, based purely on the ladder it generated, the results from Round 1 forced me to look at it again.

Just like many of us, Round 1 did not play out in line with anticipated form.

Now, of course, it’s only one game, but Round 1 got me thinking. How much stock should I place in one game?

What I’ve tried to add to the ladder algorithm is a form adjustment supplement to try and show how much my expectations should change on the basis of one game.

The predictor
First to explain the method. In short, I’ve performed a multi-linear regression to produce an algorithm that estimates how many games a team should win based on net points from the previous season, change in net points from the season before last to the previous season, and change in total wins from the season before last to the previous season.

To give an example what I’m looking at, if we take Essendon as a random example.

Points for minus points against in 2018: 94. Points for minus points against in 2018 less points for minus points against in 2017: -37. And 2018 wins less 2017 wins: 0.

(Photo by Mark Metcalfe/AFL Photos/Getty Images)

So on the basis of this model, my ladder came out like this.

Team Pred 2019
Richmond Tigers 16.25
Melbourne Demons 16.22
Geelong Cats 15.63
West Coast Eagles 14.37
Collingwood Magpies 14.21
Hawthorn Hawks 14.15
GWS Giants 13.10
North Melbourne Kangaroos 12.56
Sydney Swans 12.41
Port Adelaide Power 12.16
Essendon Bombers 11.82
Adelaide Crows 11.78
Brisbane Lions 8.71
Western Bulldogs 6.56
Fremantle Dockers 6.43
St Kilda Saints 5.98
Gold Coast Suns 2.77
Carlton Blues 2.20

It’s worth mentioning that this works purely on the correlation between the three aforementioned variables and historical win production.

I’m not taking into consideration number of injuries, who is injured, players added or deleted from the list or any other number of variables that might affect performance. It doesn’t worry about who the team has to play or strength of schedule.

In terms of performance, I applied this formula to 20 years of ladders and team wins, and the results were as follows.

Only once was my model out by more than four games in 20 years of ladders (0.35 per cent).

Only 8.83 per cent of the time it was out by between 2.5 and 3.9 games, while 44.17 per cent of the time it was out by between 1 and 2.4 games, and 46.64 per cent of the time it was out by 0 to 0.9 of a game.

In short, it was rarely very wrong and roughly 91 per cent of the time it was within 2.4 games.

My initial reaction was that this was pretty good but that’s open to speculation.

If you look at this predicted ladder, 2.4 games would make a pretty big difference in terms of ladder position for lots of teams. Positions four to 12 are separated only by 2.59 games as it is.

If we take the predicted 12th team and the predicted fourth team and apply a 2.4 game error in favour of the lower placed team, we would see a huge shift in ladder position.

Indeed, holding all things equal, such a change would take West Coast down to 11.9 wins and 11th on the ladder and boost Adelaide up to 14.18 wins, which would be good for fifth.

(Photo by Jono Searle/AFL Photos/Getty Images)

The fact that there is quite a lot of parity from fourth to 12th makes sense, and to that extent, I suppose the model isn’t ideal for predicting ladder position so much as it for predicting total wins.

Also, because teams can’t have fractions of wins, it would be necessary to round to whole figures.

Easy enough to round in line with standard ‘rounding rules’ but you could use a little of your own discretion if you had a feeling a team was better than predicted and round up from a fraction smaller than 0.5 or vice versa if you were low on that team.

Round 1 debacle
So, like many people, my tipping wasn’t great in Round 1.

I tipped in line with my expected ladder positions, which is not necessarily the way you should go about it.

It did lead me to pinch Hawthorn and Geelong against the bookies but also meant I lost big on Melbourne, West Coast, North and Sydney.

Now then, as sports media has a tendency to do where unexpected results are concerned, the reactions to these results have been pronounced.

Despite qualifying every statement with the “it’s only one game” caveat, everyone seems to have formed definite positions on how right or wrong they were only two weeks ago.

What I have tried to do, then, is apply a one-game adjustment based on the first week’s result, and Round 2 for Collingwood and Richmond.

Effectively, I have applied the difference in net points from this season and a new shift in net points based on this result. The new win totals and change in expected wins is as follows.

Team Adj Pred 1 Rd change
Richmond Tigers 16.12 -0.13
Melbourne Demons 15.85 -0.37
Geelong Cats 15.56 -0.07
Collingwood Magpies 14.62 0.40
Hawthorn Hawks 14.40 0.25
West Coast Eagles 13.84 -0.52
GWS Giants 13.80 0.70
Port Adelaide Power 12.33 0.17
Sydney Swans 12.19 -0.22
North Melbourne Kangaroos 11.80 -0.76
Adelaide Crows 11.28 -0.50
Essendon Bombers 11.14 -0.68
Brisbane Lions 9.41 0.70
Fremantle Dockers 7.24 0.80
Western Bulldogs 6.80 0.24
St Kilda Saints 6.11 0.13
Gold Coast Suns 2.83 0.06
Carlton Blues 2.02 -0.18

Again, because this system doesn’t take into account the quality of opponent, I can understand some subjective disagreement.

It’s entirely plausible that one could view a one-point victory by St Kilda over Gold Coast as detrimental to form not beneficial.

More or less, the shags in win expectation are simply in line with the size of the teams win.

Nevertheless it’s interesting to try and quantify the difference a big win or a little win makes in this regard and also to think about how that one solitary victory alters our long term projection of that team’s success or failure for the season.

If we take Fremantle, for example, my model projected about 6.5 wins. If we take my standard error this would give them a win range between about 4 and 9. Not enough to make the finals even at best and potentially a wooden spoon in the worst case scenario.

Their win on the weekend seemed to elicit responses from the general public that maybe they were wrong about Fremantle altogether.

I tend to be more conservative in this respect. That is, I tend to stick fat with the prediction, especially if I have quantitative evidence until I’m proven wrong.

Of course, this means at times I stick stubbornly to my guns while all the smart people have adjusted accordingly, but it also means I’m occasionally vindicated for staying strong while everyone reacts impulsively.

This article is not intended as an attack on Fremantle per se, they just happen to have had an unexpected result and so they are worth mentioning.

Does one win against a team many expect to be average at best mean everyone was really wrong about them?

We have all sorts of ways of justifying the position. They played with more flair and attack, or they scored more than usual.

The flip side of this is to imagine, against other more highly rated opposition, away from home no less, would we expect Darcy Tucker to have 24 touches and a goal again?

What about Cam McCarthy? Are five-goal hauls the new standard? Or is it more likely that come September we are looking back at that game as one of his few good games for the season?

And so it is with other teams. Essendon are one who seem to have evaporated any small amount of confidence anyone had in them.

My model projected Essendon as basically a 12-win team. Their hiding at the weekend makes that projection look a little closer to 11. Either way this is fairly in line with the consensus opinion on that team.

So what’s my point here? I was motivated to write something for one pertinent reason, aside from the fact that I like my little model.

AFL media is some of the most reactionary media you can consume.

One win and a team is unbeatable, one loss and the team is going to fold permanently.

The reality is, especially after the first round of the year, that these games by themselves probably don’t mean all that much. Not as it relates to the likely long run outcome anyway.

Form can often be a somewhat intangible concept and we are hardwired to place stock in the thing we saw most recently.

Sports opinion delivered daily 

   

To that end, I only urge a little perspective and patience.

Fremantle and Brisbane aren’t playing off for the flag just yet and Essendon will probably hit a target or two more over the next six months than they did last week.

The Crowd Says:

AUTHOR

2019-03-31T09:50:44+00:00

Nick Croker

Roar Guru


Nah I haven’t and that one thing could be the difference in making it more accurate. I didn’t incorporate it at this stage purely because I didn’t have time to quantify the players who moved with sufficient detail. I suppose it might be worthwhile simply giving players a value based on games played and quantifying the movement that way. For injuries getting long term data is the difficult thing. The only I can get decent data for going far enough back is the number of players that played 5 or more games each season for each team. This seems a reasonable comp for injuries but also doesn’t bother to evaluate that quality of the players injured. In any case there is a moderately strong negative correlation between having more players play 5 games or greater and winning which is what you would basically expect. The correlation wasn’t as strong as I’d expected it will give me the basis to factor in an injury adjustment as the season goes on

2019-03-31T06:56:36+00:00

Addy42

Roar Rookie


Hi Nick. This algorithm looks very interesting and will be interesting to see How it looks at the end of the season. I wonder have you made allowances for player movements and return of long term injured players. E.g. Lynch to Richmond, Dahlhaus to Geelong, loss of Lynch, May and others from Suns, return of Bruce to Saints.

AUTHOR

2019-03-30T03:38:16+00:00

Nick Croker

Roar Guru


Yes that makes sense. Not an especially good model for tipping in the short term but seems to have some accuracy over the long term

2019-03-30T03:20:53+00:00

Mister Football

Roar Guru


Your ladder prediction from the regression analysis looks surprisingly similar to what most eyeball predictions are delivering. This reminds me of a story my stats lecturer told us many years ago. She had a bet with a colleague that her predictions based on trend analysis (of either foreign currency or interest rates, or some other economic indicator) would always be more accurate than the eyeball method, i.e. zero statistical analysis. She discovered that her colleague's eyeball method delivered better predictions in the short term and that trend analaysis delivered better predictions in the long term (in other words, don't sweat the first round results).

AUTHOR

2019-03-30T03:17:15+00:00

Nick Croker

Roar Guru


Sure, all makes sense - this isn’t my permanent job so going back to find the weather conditions of every match is not something I’ve had the time to do. The accuracy and consistency of this model is pretty good for 3 simple variables.

2019-03-30T02:47:11+00:00

Doctor Rotcod

Roar Rookie


I don't think you've gone far enough. Multi-variant analysis should let you plug every score into your model,so that even though as you say there will be matches where wins and scores don't matter in terms of the season, I think that scores against the trend of play,after the siren wins, dubious umpiring and posters can all be factored in. How much weight that you give each of these might vary, given weather,i.e. heat,wind,rain or wet grounds, in-game injuries and those requiring breaks in play. Four behinds in a row might not add up to a goal, but your opponent isn't scoring, so maybe time in F50 over scores could be included. I have a background in landscape ecology where things move a lot slower but where 20 or more factors can be included in MVA

AUTHOR

2019-03-30T02:26:51+00:00

Nick Croker

Roar Guru


Yes I think you’re right - for a start I wanted to keep it simple as I could. Over time I think a trig regression would make sense and you’d see teams following a peak’trough cycle but I haven’t made it that far The other thing I like about this is that it suggests that improvement, even if not in terms of actual wins would make a difference going forward. For example if Gold Coast manages to not even win a game but we’re somehow simultaneously competitive and rarely got blown out, they could still improve their net points and trend up in terms of net point difference and have a better prognosis for 2020. The idea being that every quarter, every game matters even if it doesn’t yield a seemingly positive short term result.i.e. Reducing the margin in a dead final quarter or pinching an end of season win in a game without finals implications, blowing a team out when their spirit has been broken etc

2019-03-29T23:58:43+00:00

Gulf Drifter

Guest


I like your work Nick and agree this is a very useful - and surprisingly accurate - analytical tool. I'm going to remember this approach for Season 2020. I would reckon that an analysis of historical results would demonstrate each teams rise (or fall) up the ladder is not linear though, which suggests a non-linear or polynomial relationship is at work here as well. Perhaps other factors could be included in your formula and then a weighting applied to each factor, e.g. average games played (squad experience measure), coaching tenure (system implementation measure), ratio of injuries/playing squad in preceding years (squad factorial), level of sustained success over previous seasons (squad hunger factor), etc? Good luck with your continuing research and refinements!

Read more at The Roar