Can stats predict NRL games?

By Mike Meehall Wood / Editor

When beginning a piece that asks ‘can stats predict NRL games?’ it’s worth reviewing Betteridge’s Law of Headlines, which states that “any headline that ends in a question mark can be answered by the word ‘no’.”

If I could predict results, I probably wouldn’t be writing for a living, but that isn’t really the point here. The point is more: can stats project NRL games?

As anyone who consumes Australian advertising knows, past performance is not an indicator of future performance.

But in sports, there’s a whole category of analytics that tries to work out what aspects of performance can be relied upon to reflect future results, and indeed, which stats can be used to give a read on whether the result was actually, for want of a better phrase, the proper result.

The point is sustainability: if you won this week and the stats say you deserved it, it’s a much better read on the underlying performance than just knowing that you won. That, I would suggest, is not controversial.

The next step is to ask whether the stats that indicated that result are still valid even if you didn’t win. Logically, that should be the case too.

Lots of sports do this, but it’s been a fairly nuggety one to crack in rugby league. There’s a lot of variables at play.

Anything that involves shooting at a target tends to go alright in this regard, because you can rank how good or otherwise each shot was, and from that deduce who had the most chances to score independent of how many they did actually score.

Soccer has expected goals (xG), which measures how well you got into positions to score, and conversely, how much the opposition did that to you. ShotQuality (SQ) does something similar in basketball and there’s an AFL version too, Expected Score.

(Photo by Ezra Shaw/Getty Images)

The purpose of an xG-type stat is that it tells you who should have won, but (and any soccer analytics person should preface it with this at all times) it doesn’t really work in a one-game stat. At that level, it’s a guide and nothing more.

Where it comes into its own is as a measure of sustainability over time. The underlying numbers can prove that performances are good even when results aren’t (and vice versa), with the implication that, over a large enough sample size, they will eventually match up.

Anyone following the career of Aussie coach Ange Postecoglou might have noticed this last year: his Celtic team regularly underperformed their underlying numbers in results and rivals Rangers overperformed theirs, so at the midway point of the season, Rangers were six competition points ahead.

But the cold wind of regression waits for no man: Celtic continued to produce numbers that were good, and results began to match them, whereas Rangers continued to produce numbers that weren’t as good, and their results began to match too.

By the time the 38-game season had played out, Celtic were well clear. Celtic’s performances were sustainable and Rangers’ weren’t, just like the underlying data had suggested all along.

I’ve discussed at length the issues of creating an xG model for rugby league, as has friend of the column League Eye Test, who went as far to create a model before the publicly available stats suddenly weren’t publicly available anymore.

The short answer is that yes, getting into position to score helps, but it’s not everything. There’s plenty of ways to skin the cat in rugby league and an xG-like stat doesn’t really cut it.

Another friend of the column, PythagoNRL, used stats to work out who might have won most of the time, too, with a ‘Production Per Game’ metric that worked out what RL stats matched up well with winning and then developed probability based off that.

What we’re going to do stands in the shadow of those two. We’ve ascertained that an xG/ShotQuality model doesn’t exist in RL, but there is certainly a way of creating an xLadder – an expected ladder based on statistical production – that can help us to see performances independent of results.

(Photo by Ian MacNicol/Getty Images)

We should be able to see who is performing sustainably well, who is underperforming their output and should be expected to improve, who is overperforming output and might fall away. We can see who is Celtic, and who is Rangers.

Here’s the process. I ran a correlation analysis to discern which of all the publicly available stats are most linked to winning.

This is fairly simple stuff, because we want to cover all bases of things that win you matches of rugby league. There’s how well you attack, how well you defend, how much ball you have and where you have it. They’re the four key aspects of the game, at least at this level of modelling.

Run metres (RM) and line breaks (LB) are great indicators of attacking ability, and their negatives – run metres conceded (RMC), line breaks conceded (LBC) – tell you the opposite.

Possession is a bit more gnarly because the possession stat you see on telly is bunkum: it’s a measure of time in possession, which doesn’t really link to what you do with the ball.

Run it in slowly and take ages playing the ball and you’ll get a high possession number, but also score no points.

Instead, it makes sense to use Total Sets (TS), because (FAO Anthony Griffin) completed sets are far less an indicator of winning than simply how many attempts with the footy you had. I don’t need to go into that here other than to say that completion rate still being mentioned shows how little impact data has had on the rugby league consciousness.

Alongside that goes Tackles Inside 20m (T20), giving us an idea of field position. Note that I only include red zone tackles here, not all attacking half tackles, in order to reward ‘good ball’ sets rather than just territory.

This is a controversial move, as a perfect team would score on their first tackle within 20m, or even better, not even have one as they’d score from further out, and doubly, a perfect defensive team might be able to withstand lots of T20 from the opposition without them scoring, and have that be a positive.

Ditto a bad team (Hi, Baz!) might accumulate lots of field position without ever scoring. Still, averaged out over 16 teams and 175 games with the goal of weighting for field position, T20 is the best measure of who built the most attacking pressure across a game.

First caveat: this is model 1.0 and I fully expect to have to change this last one at a later date.

I’ve used the differentials between those four stats throughout because we want to know who won each relevant battle. Remember: teams that lose key battles statistically can still win football matches, so negatives are important there.

Next, I ran a linear regression that tests the hypothesis that these things matter and correlate to our score.

After a lot (a lot a lot) of model building all the potential combinations of these stats and others, we can say that yes, these four aspects of the game are really very related to winning.

I know you could probably have guessed as much, but hey: we’re not doing guesswork here. Double and triple check everything.

Throw in all of those stats from a decent amount of games – in this case, Round 1-10 of last year’s NRL, 80 total, and then weight the differentials to generate a score between 0 and 2 that corresponds to how likely the stats say a team was to win the game.

Put those totals in order and hey presto – we have now created an alternative NRL ladder that rewards winning the stats battle – the expected win ladder, or xLadder for short.

(Photo by Mark Kolbe/Getty Images)

Note: there were no draws in the NRL in 2022 and our model is calculated solely off full time totals, with Golden Point disregarded. This allowed for statistical comparability, and reflects how much variance is at play in extra time field goalathons.

It’s possible that future models might take this into account, but for now, we don’t. If a team wins in Golden Point in the NRL, they get two points on the real ladder, but on ours, we count differently.

The NRL ladder goes up in increments of two and ours doesn’t, because ours is designed to reflect the probability of winning based on statistical output, which can be 100% – in which case we’d award the maximum 2 points – and can be 0% – worth no points – but is almost always somewhere in between.

If the stats suggest a 50% probability of either team winning, we give a point apiece. If it’s 25%, they get 0.5, and so on.

In this lies the ability a team to under or overperform their real points on the xLadder, which should suggest where regression to the mean is likely as the sample size increase (aka the season goes on).

Guess what that means? We’re going to need a bigger sample size. I painstakingly enter the relevant statistics from every game of the 2022 NRL regular season, all 175 of them, and see where we end up.

And what does the xLadder tell us? Well, it tells us the model works, at least: we get a really clear split between the sides that finished in the top eight and those that didn’t, much wider than it was in the real ladder.

There was the Panthers, then daylight, then the Cowboys, then more daylight, then a cluster of the other finals teams, then a gaping chasm to another cluster of everyone else.

The model naturally moves teams closer together, because where the real table gives 0 or 2, it gives somewhere in between, but the spacings between the clusters are indicative.

It shows the disparity in results that, crucially match the eye test throughout. The Broncos were never as good as they looked, but benefited from a fair wind of luck, jagging several results that they didn’t really deserve. It wasn’t sustainable and regressed badly.

By the end of the year, they were about as good as the Dragons: the xLadder would have told us that in May.

Cronulla, too, prove this. They were consistently above their xPoints in real life, but when it came to finals, they bombed out in straight sets. You wouldn’t use the xLadder to pick individual games, but it certainly could have given you the steer that the Sharks weren’t quite on the level that the Cowboys and Souths were.

Parramatta, for example, showed the futility of single-game expectations, by the way, defeating North Queensland in the Prelim in a game that, on xPoints, they lost comfortably.

The Bunnies are the big winners of the xLadder. They consistently underperformed their production throughout the year and never actually caught it up – if they had, they would have made the top four. This graphic shows how they fared in comparison to the Broncos, whom they trailed for most of the year on real points, but ended up well clear of when the regression kicked in.

This speaks to two things. Firstly, it shows that their fundamentals were always good and their execution was what faltered, proving that Jason Demetriou was right to trust the process and wait for results to change. I’d be amazed if Souths don’t have some sort of modelling that suggested they were onto a winner prior to results matching the performances.

Secondly, it shows that stats are only one part of the story, because the impact of individuals is what turns theoretical wins into real ones.

The point of attacking systems in rugby league is to deliver the best quality ball to the best players at the best moment: when Latrell Mitchell returned, Souths had the man to ice the moments they had been consistently and sustainably creating all along.

The lower end of the table backs this up too. You could throw a blanket over the bottom eight, but the order is interesting.

The Knights were largely in lockstep with their xPoints until Kalyn Ponga got injured, whereupon they won just one more game despite creating enough production that, according to the model, should have delivered three wins. They were the best performing xPoints side to miss the finals, by the way.

Poor execution destroyed them, because the stats say that they were putting themselves in the position to win. I’m looking at you, Adam Clune and Anthony Milford.

Plain dumb luck comes into this as well. The Titans had five games where they ‘should’ have won based on stats, but didn’t, while also having three of their actual wins be games they ‘should’ have lost. Notably, all of those came against other rubbish teams.

Part of this is execution again, but on the other side: they continually undermined their best efforts through poor defence, even when they had ample ball, territory and opportunity to score themselves. But a substantial part is also just variance going against them.

The Dragons and Dogs make an interesting case in point, too. Prior to sacking Trent Barrett, the Dogs produced neither ladder points nor xPoints, but under Mick Potter, there was an almost immediate upsurge in performance.

They lost their first game to the Tigers and their second to Penrith but thereafter, exploded in production and eventually almost caught their xPoints.

When Barrett departed, they were a clear bottom on the xLadder, and by the end, they were only second last, but had the season continued the Dogs might have well ended up yet higher, such was their trajectory.

The Dragons, masters of efficiency, did the opposite: they bumbled along, doing about as well as could have been expected, but continued to jag results when they didn’t really deserve to. Ben Hunt executed well and they did just about enough and nothing more. It’s a skill, I suppose, but not one that I’d want to be banking on going into 2022.

This is a limitation of the xLadder, too. NRL clubs play 24 regular season games a year, which is a pitifully small sample size compared to the 38 in the Premier League, 82 in the NBA and NHL and 162 in Major League Baseball.

The potential for regression increases as the sample size grows, so the NRL season can run out before the results catch up with performances. If you’re Kevin Walters or Anthony Griffin, that should be worrying, but if you’re Adam O’Brien or Jason Demetriou, it might feel alright.

We should also acknowledge the general lack of sample size: the rules within the NRL have changed so much in the last two years that any data prior to 2022 is functionally not that useful. 2021 was the asterisk year, 2020 was half a season and 2019 was totally different rules.

That brings us back to predictions and probabilities. I’ve set out that there is a clear link between underlying data and results, and how that can be modelled.

I’ve then built the model, explained it (I hope) and showed how, using last year, it identified several key things that did actually occur. The next step is 2023 and looking into how it tells us the season will unfold. After all – we might not be able to predict the NRL, but we might be able to project it.

The Crowd Says:

2023-02-16T08:44:20+00:00

mushi

Roar Guru


It's more the other way round using stats to attribute the actual wins to individuals rather forecast team wins. There are expected points models in basketball but you find they better for defence as good offense is about out performing. The stats principles in football are a bit different because of the rarity of scoring. Basketball on a team basis is typically done on pure points allowed/scored per possession

2023-02-15T03:28:32+00:00

Hondo

Roar Rookie


Thanks Mike appreciate your efforts my only disagreement was your comment about them "because their performances were quite consistently not that good throughout" as i thought on that night they were very impressive, sadly for them they couldn't reproduce the effort and commitment that they showed that night in the first half. They really had me thinking they were a very good team that night and a class above Parramatta, which fortunately for the Eels didn't turn out that way as the season progressed. I really think they have the players to perform at a higher level this season. I have my fingers crossed for the Eels but believe they won't make the eight this year without the players they lost on the off season.

2023-02-14T07:07:41+00:00

Short Memory

Roar Rookie


And you need to use the word "Shock" more frequently in your headlines. :silly:

2023-02-14T06:03:01+00:00

blacktown leagues

Roar Rookie


my head hurts

2023-02-14T05:47:16+00:00

mushi

Roar Guru


He can rest when it's been tested!

2023-02-14T05:15:39+00:00

Brett Allen

Roar Rookie


Yeah the NBA uses Win Shares, Offensive Win Shares & Defensive Win Shares. Basically it calculates how much each player contributes to his teams wins. For example Kareem Abdul-Jabbar has the most Total Win Shares with 273.41, LeBron is second with 254.52, MJ is 214.02. In terms of Offensive WS, Kareem is 1st with 178.92 Defensive WS, Bill Russell 1st with 133.64, Tim Duncan 2nd with 106.34. The only two with more than 100 career Defensive Win Shares.

AUTHOR

2023-02-14T04:24:10+00:00

Mike Meehall Wood

Editor


I looked it up and the specific 'jammy gets' example was actually me calling the Cowboys lucky in a Power Rankings where I had the Broncos 2nd and called them 'the real deal' :laughing: I also came (i think) dead last in The Roar tipping comp...

2023-02-14T04:22:19+00:00

Albo

Roar Rookie


But I don't need to look at set of stats to tell me the Dragons & Titans are inferior teams to the Storm . Nor do I need a set of stats to tell me that Souths have better cattle than the Broncos at this time . As far as I am concerned the better teams will generally have better stats with all other aspects being equal ( injury rates, refs calls, bounces of the ball, etc)

AUTHOR

2023-02-14T04:22:03+00:00

Mike Meehall Wood

Editor


So there's two points here: 1 - As mentioned in the piece, single game examples aren't really what this is about. But, since you asked...the Broncos lost the xPts that night meaning they *should* have lost. But, again: I went and watched the highlights there and they were little lucky (good bounces, sin bins, Parra brainfades) but also maximised opportunities, which is in itself a good thing thata other sides didn't do. Games like that are why it doesn't really work as a single game example, but anything can happen. The point is to see trends over time. 2 - Secondly, that game is a good example of another concept, which I'll get into later in the season: "game state". The Broncos blitzed Parra before halftime, so it didn't matter that the Eels ran up lots of stats in the second half because all the Bronx had to do was stop them scoring, they didn't need to post their own offence. That would impact the xPts on a single game basis and make the game seem closer than it was. It's defo a limitation of the model but one that, currently, I can't factor in. I went and checked, and at HT, the Broncos comfortably led all metrics, so the game state allowed them to play totally differently in the second 40. I'm going to track this from the start of this season to see how it can improve the model going forward.

2023-02-14T03:34:21+00:00

Pomoz

Roar Rookie


Great article Mike, I love your work. I see you have ignored my career advice to write articles sprinkled with hyperbole, teeming with unsubstantiated facts and rumours and concluding with the premise that it is probably all Gus Gould's fault. Oh and I nearly forgot, at least one paragraph of stating the bleeding obvious (you know "SOO will be tough" this year, that sort of thing. Or my favourite "they need to get the ball to their fast players in the backs" an Anasta special). See what all this thoughtful, insightful analysis has done for you? You still haven't appeared on NRL 360 or NRL tonight or the Matty Johns Show. I tried to help........

2023-02-14T03:21:38+00:00

Pomoz

Roar Rookie


Well stats are a measurement of the output of a team, so yeah, good teams produce good stats. Of course lady luck has a role to play. Two minutes in and your halfback and five eighth head clash each other and get stretchered off. Your team may have the stats in the entire NRL but lady luck has just thrown that out the window. Fast rewind Penrith vs Eels in 2022, Cleary is sent off and the centre gets injured. During the ensuing confusion in defence, ten minutes later the Eels have iced the game. Stats be damned. But this is much better than just looking at the cattle because it allows you to really understand if a win is lady luck, or inline with their usual performance. It may be easy to look at the cattle and say "good team" when you look at the Storm, but what about the Dragons or the Titans? What is actually happening underneath the hood? Can the stats show where they will finish and explain why they lose? It is a very interesting revelation that South's underperformed against their stats during the first part of the season and the Broncos over performed. What this tells you is that if South's can maintain the level for a whole season, they might go all the way. The Broncos on the other hand need to change what they are doing or they will not make the eight again. If they perform to the stats level for the whole season and don't get any luck they might finish lower than 2022.

2023-02-14T03:03:14+00:00

The Barry

Roar Guru


They were heady days….

2023-02-14T02:55:11+00:00

Forty Twenty

Roar Rookie


''Unlucky'' and poor recruitment and retention has been the number one issue at the club. Ultimately it has also been Manly's big issue for a decade now. When the dust has settled if the players you've recruited haven't done as well as the players you've let go then that's the issue. Plenty of good cattle at both clubs now but if the young half the Broncos couldn't keep, leads the Berries to the crown sometime, and the Broncos miss out then it's history repeating.

2023-02-14T02:43:02+00:00

Forty Twenty

Roar Rookie


It worked well at the start but even as a rival fan who lost Des as a coach and a few top players I felt sorry watching them play in his final year or two. Des got it wrong at the end , no doubt about it but he did very well for much of his time at the club.

2023-02-14T02:36:28+00:00

Heyou

Roar Rookie


I haven’t read it yet but I’m going to. Feeling at death’s door atm. I’ve saved it. It looks pretty special.

2023-02-14T02:33:18+00:00

Heyou

Roar Rookie


:laughing: :laughing: :laughing:

2023-02-14T02:30:57+00:00

matth

Roar Guru


"has nothing to do with statistics but a lot to do with the effort each team put into each of the respective games". I think you will find that the effort put in plus skills, produces the statistics

2023-02-14T02:24:44+00:00

Hondo

Roar Rookie


Im interested in the analysis of the first Broncos v Parramatta game last year when the Broncos performance was a class above Parramatta both by backs and forwards. Watching that game there was no doubt who was the better team that day as The Broncos destroyed Parramatta. I have no idea of statistics but the relative performances of Parramatta and the Broncos in the first game as against the total domination by Parramatta over the Broncos in the second to me has nothing to do with statistics but a lot to do with the effort each team put into each of the respective games. So what changed as while the Broncos declined in effort late in the season the Eels definitely lifted their effort towards the end of the season. So i cant agree with that comment "They didn’t ‘slump’ in performances, only in results, because their performances were quite consistently not that good throughout" as my eyes tell me differently on the basis of their on field performance in both games.

2023-02-14T02:23:58+00:00

matth

Roar Guru


I remember and I remember both resenting it and at the same time being fearful you were correct, which you were.

2023-02-14T02:08:55+00:00

Adam

Roar Guru


WARG sounds particularly good and I'll have to take a look at the pythagoNRL. I'm assuming something like that could be accounted for across positions as long as the fundamentals are weighted correctly across the population. And I would say it's something similar with the CV that expands and makes allowances for rounded halves such as Cleary who will force more drop outs or positional pressure as opposed to Ilias who will score more points but won't apply the same pressure points. I'd say if there are higher value being attributed to Cleary that's probably on the money

More Comments on The Roar

Read more at The Roar