AFL Prediction & Analysis

The Squigglies 2022: Ladder Predictions

Oh sure, now, everyone looks back on the preseason ladders and mocks how wrong they were. “Essendon to make finals,” they say, shaking their heads. “Not even close.”

But no-one was close, of course; everyone’s ladder has a howler or two. If you picked Essendon to fall, you probably didn’t also pick Collingwood to rise, or Port Adelaide to miss.

That doesn’t mean they’re all equally bad, though. Here at Squiggle, we value the signal in the noise, even if there’s still a lot of noise. And ladder predictions that were less wrong than everyone else’s are to be celebrated.

Every 2022 Expert Ladder Prediction Rated

Best Ladder: Peter Ryan

This is a heck of a good one, and it’s no flash in the pan:

Ryan’s ladder managed to get 7/8 finalists, which is fantastic given that three of them finished last year in 11th, 12th, and 17th. (His tip of Fremantle for 6th — a single rung too low — was especially good.) Like everyone else, he missed Collingwood, but correctly foresaw exits by Port Adelaide, Essendon and GWS. He also resisted the popular urge to push Geelong down the ladder, and wisely slotted the Eagles into the bottom 4.

Damian Barrett also registered a good ladder this year, with 6/8 finalists and three teams in the exact right spot. There was a fair gap from these two to Jake Niall in third.

Runner-Up: Damian Barrett

Best Ladder by a Model: Squiggle (6th overall)

Squiggle nudged out other models with some optimism on Sydney and pessimism on Port Adelaide, but not enough of the former on Collingwood and not enough of the latter on GWS and the Bulldogs.

Honourable Mention: The Cruncher (11th overall)

Long-Term Performance Award: Peter Ryan

Not everyone publishes a ladder prediction every year — it’s a little shocking how frequently journalists come and go from the industry — so although I always have a bag of 40 or 50 experts and models to rank, only half appear in all four of the years I’ve been doing this. Of those, Peter Ryan has the best record, finishing 19th (out of 45), 9th (/56), 3rd (/42) and 1st (/45). That’s an average rank of 8th, making him the only one to outperform Squiggle over the same period.

Honourable Mention: Squiggle (5th, 20th, 9th, 6th)

Live Running Predictions

Squiggle pipped AFLalytics and Wheelo Ratings on the Ladder Scoreboard this year, mostly thanks to some solid returns in the early rounds.

Throughout the year — but especially early — the teams models overrated the most were GWS and Hawthorn, while they underrated Collingwood and Fremantle.

Introducing Power Rankings

Last week, in the Squiggle models group chat – of course there’s a group chat – Rory had a good idea:

Rory’s idea

It turned out that everybody had data on hand for this, because if you have a model, you also have a rating system. So I began collecting this, and now there’s a page to view it.

There’s also a widget here on the site, to the right of this post, or else above it.

On the main page, you can see how ratings change over time, and compare ratings from different models.

Power Rankings measure team strength at a point in time. They ignore the fixture, home ground advantage, and all the other factors that go into predicting the outcome of a match or a season. Instead, they’re a simple answer to Rory’s question: Which teams are actually good?

The Curse of the Curse

I enjoy a useless AFL stat as much as the next person, but this kind of thing tests me:

“Curse” is a bit of a tell in footy. It usually means “coincidence.” If it was a real effect, we’d have a decent theory about why. People love to invent theories. There’s no effect we won’t try to pair with a cause, no matter how thin the evidence. When there’s an effect and no cause, I tend to doubt it’s due to the spooky unseen hand of an unnamed force.

Usually a “curse” is an odd stat that, at first glance, seems like it can’t be the result of random chance, but that’s only because we don’t understand randomness. Our gut tells us that flipping five heads in a row is basically impossible, for example, when in fact true randomness tends to contain a lot more natural variation than people think. (You can flip ten heads in a row, if you’re willing to toss coins for a few hours, and people will think you’re a magician.)

Here’s the 0-2 stat:

I have a few problems with this.

First, I have to point out it’s technically wrong, because we’ve had nine finalists from 0-2, counting Carlton in 2013 who were elevated from ninth after Essendon’s disqualification.

But more importantly, the underlying effect sounds suspiciously like “It’s harder to make finals if you lose games.” And we knew that already. Is there anything magical about the first two games? Because if not, it’s just saying that dropping games hurts your finals chances.

Then there’s two snipes: the starting point (2010), and the number of games (2). If there’s a genuinely interesting effect here, and not a coincidence, we should expect to see not-quite-as-dramatic-but-still-suggestive numbers when those key numbers are varied a little.

Instead, it vanishes pretty abruptly. If you look at a longer time period, you see about 20% of 0-2 teams making finals, and if you look at 0-1 or 0-3 or 0-4 teams, the numbers again are about what you’d expect: about one-third of 0-1 teams make it, about one-in-ten 0-3 teams, and only Sydney 2017 has made it from 0-4 this century. So the more games you lose, the harder it is to make finals, in a steady and predictable way.

Because what actually happened here – the whole reason this stat became popular – is that between 2008 and 2016, there was a patch where only two 0-2 teams made finals (Carlton 2013 and Sydney 2014). This hit rate was quite a bit lower than the years before and after, although not wildly so:

YearFinalists from 0-2
20002 out of 5
20011 out of 5
20021 out of 5
20031 out of 3
20042 out of 4
20050 out of 3
20062 out of 3
20071 out of 4
20080 out of 4
20090 out of 4
20100 out of 6
20110 out of 4
20120 out of 5
20131 out of 7
20141 out of 6
20150 out of 5
20160 out of 4
20171 out of 8
20181 out of 4
20191 out of 5
20201 out of 4
20213 out of 5

Eyeballing that, you might notice something else about the middle years: There are more 0-2 teams. And indeed we had a number of clubs at historical lows in this period, including two teams who were introduced to the league. Fourteen of those 0-2 non-finalists from 2008-2016 are actually just four clubs failing over and over: the two expansion teams plus Melbourne and Richmond.

So this always looked a fair bit like random variation plus an unusually weak bottom end of the comp. But somehow it gave birth to a “curse” that meant flag contenders couldn’t afford to drop their second game.

And now that regular service has resumed – implying that there was never much to see in the first place – “a new trend is emerging.”

Ladders of Future Past

You can now use the ladder predictor on seasons as far back as 2000. Relatedly, the Squiggle API now serves fixture info on games dating back to 2000, and you can also use it to get a list of which teams were playing in any of those years.

You might be wondering why you’d ever want to predict past ladders. To be honest, I’m not sure. I just know that people write in sometimes asking if the site can let them do that.

This particular addition was triggered by Jake, who emailed me to say he’d been in iso for a month, and he kept busy by re-entering past seasons into the predictor one game at a time to see how the ladder changed. Jake had done this for 2011-2022, but wanted to go back further.

So now you can. I am all about football as a mental escape from reality, Jake. That’s the best possible use of football.

The Squigglies 2021: Pre-Season Ladders

Heading into 2021, there was a bit of hive mind syndrome going around:

So everybody had Richmond way too high, and Melbourne, Sydney and Essendon too low. Collingwood were generally tipped for somewhere around mid-table, often pushing into the Eight, as were St Kilda.

This same-same field of predictions delivered neither a spectacularly good nor spectacularly bad ladder. Instead, everyone was just kind of okay. The average was better than just tipping a repeat of 2020, but not by much.

Every 2021 Expert Preseason Ladder Rated

Best Ladder: Daniel Cherny

All year long, the Western Bulldogs looked a deserving top 2 team. Then they plunged from 1st to 5th in the final three rounds, upending a lot of ladder predictions along the way. A benefactor was Daniel Cherny, who’d tipped them for 6th, and suddenly had the best projection out of anyone. He had 6 of the Top 8, missing Sydney & Essendon for Richmond & St Kilda, and half the Top 4. He also wisely tipped Collingwood to fall further than most (although not as far as they actually did).

Runner-Up: Sarah Black

Best Ladder by a Model: The Flag (6th overall)

After coming second in this category last year, this was a great performance by The Flag, nailing three out of the Top 4, with Richmond the only miss.

Honourable Mention: AFLalytics (8th overall)

Lifetime Achievement Award: Peter Ryan

Of the 26 experts and models I’ve tracked for three consecutive years, Peter has the best record, averaging 65.03 points across that period. He’s been getting better, too, finishing 19th in 2019, 9th in 2020, and 3rd this year.

Honourable Mention: Squiggle (5th in 2019, 20th in 2020, 9th in 2021)

Mid-Season Predictions

If you’re interested in how models predicted the final ladder during the season, head on over to the Ladder Scoreboard. New model Glicko Ratings scored best this year, while as usual all models significantly outperformed the actual ladder.

Ninety-nine percent

via maxbarry.com

If you do one thing each day that has a 99% survival rate, you’ll likely be dead in under ten weeks. If boarding a plane had a 99% survival rate, a typical flight would end by carting off at least one passenger in a body bag, perhaps two or three. Ninety-nine sounds close enough to 100, but anything with a 99% survival rate is incomprehensibly dangerous.

Go sky-diving, and you’re over two thousand times safer than if you were doing something with a 99% survival rate. Driving, the most dangerous everyday activity, requires you to clock up almost a million miles of travel before you’re only 99% likely to survive. Even base jumping, perhaps the single most dangerous thing you can do without actively wanting to die, is twenty-five times safer than anything that carries a 99% survival rate.

Ninety-nine bananas is essentially one hundred bananas. Ninety-nine days is practically a hundred days. But 99% is often not even remotely close to 100%. It feels like similar numbers should lead to similar outcomes, but the difference in life expectancy between 99% and 100% survivable daily routines isn’t one percent: It’s ten weeks versus immortality.

It’s simple enough to calculate the probability of more than one thing happening: You just multiply the individual probabilities together. The likelihood of surviving for three days, for example, while doing one thing per day with a 99% survival rate, is 0.99 x 0.99 x 0.99 = 0.9703, or 97.03%.

But we find this deeply counter-intuitive. We prefer to think in categories, where everything can be labeled: good or bad, safe or dangerous, likely or unlikely. If we have an appointment and need to catch both a train and a bus, each of which have a 70% chance of running on time, we tend to consider both events as likely, and therefore conclude that we’ll make it. The actual likelihood that both services run on time is 0.70 x 0.70 = 0.49, or only 49%: We’ll probably be late.

We also prioritize feelings over numbers. Here’s a game: Pick a number between 1 and 100, and I’ll try to guess it. If I’m wrong, I’ll give you a million dollars. If I’m right, I’ll shoot you dead. Would you like to play?*

Most people won’t play this game, because the thought of being shot dead is too scary. It’s shocking and visceral, so when you weigh up the decision, both potential outcomes balloon in your mind until they feel roughly equal, as if the odds were 50/50, rather than one being 99 times more likely than the other.

But put the same game in a mundane context — if instead of being shot, you get COVID, and instead of a million dollars, you just go to work as usual — and we tend to return to categorical thinking, where the dangerous-but-unlikely outcome is filed away as too improbable to be worth thinking about. As if close to 100% is close enough.

Between 99% and 100% lies infinity. It spans the distance between something that happens half a dozen times a year and something that hasn’t happened once in the history of the universe. With each step we take beyond 99%, we cover less distance than before: 1-in-200 gets us to 99.50%, then 1-in-300 to 99.67%, then 1-in-400 only to 99.75%. We’ve quadrupled our steps, but only covered three-quarters of the remaining distance. We can keep forging ahead forever, to 1-in-a-thousand and 1-in-a-million and beyond, and still there will be an endless ocean between us and 100%.

You have to watch out for 99%. You have to respect the territory it conceals.

* I pick 73.

Ladder Prediction 2021

You know what, too many people are doing half-arsed ladder predictions. By which I mean, they’ll only tip the top 8, or give a range of possible finishing values, or say who will rise and who will fall but not by how much.

That’s garbage, people. Yes, it’s difficult. Sure, nobody will ever get it just right. You can still have a crack, and let me measure it.

Here is Squiggle’s prediction for 2021. No really hot takes this year, and it’s going to be a tough one after an unusual 2020. But this is the model’s attempt after factoring in off-season movements, long-term injuries, and preseason form (yes, that one practice match).