The Squiggle stable of models has grown to ten this year, and now includes almost all of the well-known public tipping models from across the web.
This site aims to bring together and promote the best of such models, and I’m interested in adding more. In order to keep the quality of such models high, this is what I look for:
- An official site where tips are posted
- A history of doing so for at least a year
- Public discussion and analysis of football
- Some transparency about the model
These aren’t all mandatory. There are currently some very mysterious models included in the stable because they are widely known and respected. But this is the ideal.
I know there are many solid tipsters with an Excel spreadsheet and a Twitter handle, which I don’t plan to list since they don’t meet the above criteria.
Ultimately my goal is to expose more insight into the how and why of football analysis, looking not only at who is tipping best, but why. What is the best way to treat Home Ground Advantage? How much difference does the weather make? How important is it to consider player rankings, or recent form? What factors matter most in determining which side wins a game of football?
The official AFL ladder predictor has a few problems:
- Not available until late in the season
- Requires ten thousand clicks to complete
- Created by Telstra monkeys in 1994
So I made a new one!
Now you can tip as few or as many games as you like and lean on the world’s best computer models to fill in the rest.
You can even go back and change some of the computer tips if you want.
I believe the world is a better place when people can generate wildly optimistic ladder projections with ease. At the moment it’s only for the home & away season, but I’ll probably add finals sometime later.
How AutoTip Works
The Predictor fetches data from the Squiggle API, including Aggregate tips from computer models such as The Arc, Matter of Stats, FMI, HPN and many more. This provides a tip for each match as well as a confidence rating about how likely the team is to win.
It’s no good to simply tip the favourite in each game, though, because in real life, favourites don’t always win. For example, if two people play 10 games of chess, and one player is 60% likely to win each match, we want to be able to predict that the final tally will be about 6-4, and not 10-0, like we’d get if we tipped the favourite to win each time.
So AutoTip runs several thousand simulations, each time applying an amount of randomness to the predicted result, so that a team that is 70% likely to win a match will only win it in about 70% of simulations.
The simulations are then analyzed to determine the average finishing rank, number of premiership points, and percentage of each team.
This provides a pretty reliable estimation of where each team is likely to wind up on the ladder. However, we don’t merely want average numbers: We want to see specific tips for each game. So next AutoTip scores each simulation based on how closely its ladder resembles the average. Then it selects the “most normal” one.
This means AutoTip will contain many upsets, and which upsets they are will change each time. But the upsets will be spread around evenly, so that each team finishes the season about where they’re expected.
Richmond have been doing two things Squiggle particularly likes: Holding oppositions to low scores, and generating plenty of scoring shots.
The Tigers have been kicking plenty of behinds this season right from the start: In Round 1, they defeated Carlton by 4 goals and 14 behinds, while in losing to Adelaide the following round, they lost by 6 goals and 0 behinds.
On the surface, that’s a close-ish 26 point win followed by a heavier 36-point defeat; in terms of scoring shots, it’s a +16 smashing followed by a much closer -6.
Squiggle’s model considers the reality to be somewhere in between. As a result, it considers the Tigers’ only loss of the season so far to be a relatively close one, away interstate to a very good team – the kind of game that even a top team will often drop. The Tigers’ wins, on the other hand, have included some extraordinary smashings when viewed in terms of scoring shots.
Richmond’s opposition to date has mostly included mid- to upper-tier rated teams in Adelaide, Collingwood, Hawthorn, and Melbourne, yet across the season’s 6 rounds, the Tigers have averaged 50% more scoring shots. The result is a lot of Squiggle love for the yellow and black.
The Tigers have a monstrous gap over the rest of the comp, according to Squiggle modelling. This is the biggest gap between the best and second-best sides since the mid-2012 Hawks. pic.twitter.com/JbGufTJdtv
— Squiggle (@SquiggleAFL) April 30, 2018
Most computer models of AFL have no awareness of who’s actually playing in either team. They rate the strength of competing clubs as an entity, rather than what they really are: a collection of moving parts, some of whom move a lot better than others.
Earlier this year, HPN, the artist formerly known as Hurling People Now, launched PERT, a tipping model that draws on the ratings of the actual players, as determined by the PAV system.
It’s early days, but PERT looks likely to add a fascinating new dimension to model-based tipping, and is definitely worth keeping an eye on.
Because PERT tips depend on the announced teams, they come out later in the week and can alter based on last-minute personnel changes.
I’m delighted to have HPN on board and PERT tips are now integrated into Squiggle dials and the leaderboard!
People can build great things in AFL analytics if we share data! Here’s a first draft of Squiggle’s API, which provides machine-friendly access to game & tips data:
The man behind the model, Professor Stephen Clarke, has kindly agreed to join in, and Swinburne is now included in the models leaderboard and tips pages.
Due to confidentiality requirements, Swinburne margin and probability tips are embargoed until the completion of each game: until that point, only the name of the tipped team is displayed.
It’s very exciting to have the great grandaddy of AFL computer models on board.
Swinburne is off to a solid start this year, too, shooting straight to #1 on the Squiggle leaderboard! A good challenge for the rest of us.
Exciting times with two new models added to the Squiggle dials:
- Ranking Software, also known as AFL Live Ladder , also known as Darren O’Shaughnessy, one of the absolute doyens of AFL analytics
- Stattraction, also known as statisticc from Reddit’s r/afl
And that’s not all! Two more big ones coming next week.
I’ve now fixed the code that dumbly extrapolated out scoring shots of live matches, the problems with which were kindly highlighted by North Melbourne and St Kilda last week.
— Squiggle (@SquiggleAFL) March 30, 2018
Squiggle would now project those 2nd-quarter scores of 2.9 (21) – 2.7 (19) out to 8.16 (64) – 10.15 (75)… which is more plausible. Although still not very close to the actual final scores of 13.17 (95) – 5.13 (43).
Got some more updates ahead of season 2018:
- Pre-season Tracking: We had something resembling a normal preseason this year, so Squiggle is using those results in its calculations for the season ahead.
- Round-based Sensitivity: Squiggle is now more sensitive in the early rounds and in finals, to better capture the important signalling of surprising results at these times. In particular, an unexpected Round 1 result can cause a team to move a very long way.
- Goal-kicking Accuracy: As flagged last year, since goalkicking accuracy seems to be non-reproducible — teams that are highly accurate one week aren’t more likely to be highly accurate the next — Squiggle will interpret a scoreline like 13.20 as a team that was unlucky to not win by more, and thus will rate them more highly, while a scoreline like 12.3 will be interpreted as a team that got closer than they probably deserved.
- Home Ground Advantage: Gone is the practice of assigning 12 points of home advantage to interstate hosts and ignoring everything else! Squiggle now calculates HGA based on relative ground & state familiarity: It scores each team on how many times they’ve played at the same venue and (less significantly) in the same state over recent years, and assigns HGA points based on how lopsided the numbers are. So at venues where both teams have played fairly often, ground familiarity is balanced and there’s not much HGA, if any. At venues where one teams plays a lot and the other team rarely even visits the state, HGA is high.
The main difference in practice is better modelling of Geelong in Melbourne, along with uncommon situations like Melbourne playing in the Northern Territory. For example, for the Round 1 Melbourne vs Geelong game at the MCG, under the old model Squiggle would give Melbourne 12 points of HGA, but under the new model, it’s only 2.5 points, due to the Cats’ fairly good familiarity with that ground.
Actually here are the current HGA values for all R1 matches:
- Richmond +1.1 vs Carlton @ MCG
- Essendon +7.9 vs Adelaide @ Docklands
- St. Kilda +7.6 vs Brisbane @ Docklands
- Port Adelaide +10.0 vs Fremantle @ Adelaide Oval
- Gold Coast +10.3 vs North Melbourne @ Cazalys
- Hawthorn -1.0 vs Collingwood @ MCG
- GWS +9.0 vs Bulldogs @ UNSW Canberra
- Melbourne +2.5 vs Geelong @ MCG
- West Coast +10.9 vs Sydney @ Perth Stadium
Hawthorn are the only home team to have negative HGA. West Coast still get a bunch of HGA despite playing on a brand new ground due to the state-based numbers.