Research & Methodology
How Accurate Are Property Forecasts?We Built One. Tested It. Abandoned It.
There’s no shortage of property forecasts in Australia. Every major bank, every research house, every buyer’s agent publishes growth predictions. Almost none of them publish how accurate their previous forecasts actually were.
We do — because we tested ours, and the results forced us to change course. We built a suburb growth forecaster. It looked compelling. Then we ran a test that most forecasters never bother with, and what we found changed everything about how BoomAU works.
Two Questions That Sound the Same
The distinction that rarely gets made in property commentary, but matters more than almost anything else:
Prediction
Which suburbs will grow in the next 12–24 months, before they start?
Requires leading indicators. Sounds like exactly what you want. Failed at 55% accuracy — a coin flip.
Detection
Which suburbs are currently in a growth phase, and how far through are they?
Requires real-time market signals. Sounds like you’re already late. Hit 85.7% accuracy.
These look like variations on the same question. The data requirements are completely different, and the accuracy gap is enormous. Here’s what the testing showed.
The Prediction Formula: 55%
The first formula used the metrics that property commentators reach for automatically: infrastructure spending, population growth, building approvals. The theory was that these leading indicators could identify boom suburbs 12–18 months before the market caught on.
Prediction formula — inputs
- •Infrastructure spending near the suburb
- •Population growth and migration inflows
- •Building approvals (supply pipeline)
- •Rental yield
- •Vacancy rate
- •Vendor discount
- •Days on market
A coin flip. With seven inputs and a lot of extra work.
The core problem: infrastructure projects affect whole corridors over decades, not individual suburbs over 18-month windows. Population growth data moves too slowly and is too coarse to rank suburb A against suburb B. Building approvals are measured at state level — they tell you about supply pressure in a broad area, not whether a specific suburb’s buyers are running out of options. Vendor discount data barely exists at suburb level from free sources.
None of these signals have the resolution needed to rank individual suburbs against each other over an investment-relevant time horizon. They’re meaningful context for understanding broad market direction. They have near-zero predictive power at the suburb level.
Key finding
Infrastructure spending, population growth, and building approvals failed as suburb-level predictors. The metrics move at city or corridor scale. A suburb-level decision requires suburb-level signals.
We only score using signals that survived backtesting.
393 suburbs, fortnightly. No prediction. Just detection. Join the wishlist.
The Forecaster That Looked Convincing
After the prediction formula failed, we tried something more sophisticated: a proper suburb forecaster. A model that would predict 3-year forward capital growth for every suburb, with confidence ranges. Six inputs: annual growth rate, how affordable the suburb was relative to its city median, whether a boom had already been detected, 5-year growth momentum, the national market cycle, and the recent acceleration in growth.
We ran it across thousands of suburb-months of historical data. The headline result was compelling.
A rank correlation of 0.42 is a strong signal. It suggested the model was genuinely sorting suburbs in the right order. Most quantitative approaches to property are happy to find any consistent signal at all.
For a moment, this looked like a working forecaster. Then we ran the test that changes everything.
The Test That Killed It
Most property forecasts are evaluated by pooling all their predictions together and measuring how well they correlated with outcomes overall. That feels rigorous. It’s not.
Here’s the problem: when you’re deciding where to buy, you’re making that decision in a specific month. You are not choosing between a suburb available in 2017 and one available in 2023. You’re choosing between two suburbs that are both available to you right now. The question the model needs to answer is: within this month, which suburb will outperform?
So we split the analysis within each individual time period and asked: does the model’s ranking of suburbs predict their relative performance in that same period?
Negative 0.08. Worse than random.
The pooled 0.42 was a statistical illusion. Boom years produced high growth broadly — the market tide lifted every suburb. Our model happened to give higher scores in boom years. So the pooled correlation looked strong. But within any given month, the model had no ability to tell you which suburb would beat the others. It was ranking time periods, not suburbs.
The confidence ranges were just as broken. When a forecast claims 80% confidence — meaning the real outcome should fall within the predicted range 80% of the time — it needs to actually hit that 80%. Ours covered 46%. At the moments the model was most confident, coverage fell to 20%. The more certain the model was, the more wrong its confidence was.
We abandoned the forecaster entirely.
Key finding
The tide lifts all boats. Growth phase, momentum, and acceleration do not predict which suburb outperforms within any given period — they track the overall market cycle. A pooled correlation that looks strong can mask a within-period correlation that is worse than random. Always test both.
Prediction: 55%. Detection: 85.7%. We use detection.
Join the wishlist and see which suburbs our backtested formula is watching.
Why Detection Works Where Prediction Doesn’t
The insight that changed the approach: property booms are not events you catch before they start. They are multi-year processes you can catch early.
A suburb boom that runs for three years, detected six months in, still has the bulk of its growth ahead. The numbers from the backtest: entering a detected boom 6–12 months after it starts still captures 60–85% of total gains. The early-mover advantage of perfect prediction is much smaller than it looks — and the accuracy cost of trying to predict is enormous.
Boom size is also era-dependent: the median boom in the pre-2015 data was 1.3% growth above trend. Post-2020, that figure was 16.2%. In high-momentum periods, the cost of arriving late is lower than the cost of getting the wrong suburb entirely.
Detection asks a different and more tractable question: is this suburb booming right now, and how far through the boom is it? That question can be answered with current market data — price growth, days on market, vacancy rate, rental yield, and affordability relative to the city median. All publicly available. All suburb-specific.
The Detection Results
The detection formula was validated against 78 suburbs — 28 that boomed, 50 controls that didn’t. Hard filters first: any suburb with annual growth below 5%, days on market above 45, vacancy above 2%, or a median price above $800K was excluded from scoring. Within the qualifying pool, five components combine into a score.
Detection formula — 5 components
- ✓Momentum (30%) — Price growth acceleration — is the market speeding up or slowing down?
- ✓Growth Strength (25%) — Annual growth scored directly against thresholds
- ✓Tightness (20%) — Days on market combined with vacancy rate
- ✓Sustainability (15%) — Rental yield and vacancy trend direction
- ✓Headroom (10%) — Suburb median price relative to the capital city median
The separation gap matters as much as the accuracy number. Real booms score 20.2 points higher than false signals on average. The formula doesn’t just get the right answer — it gets it with margin. These aren’t borderline calls.
Zero false positives means no suburb the formula flagged as booming turned out to be a non-event. The accuracy drops are on the other side: the formula occasionally misses a boom that didn’t exhibit the expected market tightness signals. 85.7% leaves some booms undetected — it doesn’t generate ghost signals.
Does the Scoring Actually Discriminate?
Detecting booms is one thing. The harder question: among all scored suburbs, do higher-scoring suburbs actually outperform lower-scoring ones? A walk-forward backtest across 12,360 postcode-months answers this.
| Tier | Excess return | Beat market | n |
|---|---|---|---|
| Strong Buy | +7.5pp | 71% | 2,103 |
| Buy | +1.3pp | 55% | 3,349 |
| Watch | −0.7pp | 47% | 5,788 |
| Pass | −6.4pp | 28% | 1,120 |
Walk-forward backtest, 12,360 postcode-months. No lookahead. Excess return = suburb 12-month growth minus market median growth. Full methodology →
Perfectly monotonic across all four tiers. The top tier outperforms the market by 7.5 percentage points and beats it 71% of the time. The bottom tier underperforms by 6.4 points and beats it only 28% of the time. The spread from top to bottom is 13.9 percentage points.
This is the discrimination the forecaster couldn’t produce. Not which suburbs will grow in absolute terms — but which suburbs will outperform their peers within the same period.
The Two Signals Behind the Accuracy
Of all the signals tested, two survived every subsample and every version of the formula:
1. Affordability headroom
A suburb’s median price relative to its capital city median. Every boom in the 78-suburb backtest was led by a suburb priced well below the city median. Suburbs priced above 1.5x the city median consistently underperformed. This is the only cross-suburb ranking signal that survived tide cancellation — it works because demand displacement flows to where buyers can still afford to buy.
Check it: Domain publishes city medians quarterly. Compare the suburb’s CoreLogic median (via YIP) to the capital city figure.
2. Boom timing via market tightness
Is the suburb in a detected boom, and how early is it? Annual growth above 5%, days on market below 45, and vacancy below 2% together define the boom signature. The detection formula also measures how much of the suburb’s affordability gap has already closed — a suburb where less than 30% of its headroom has been consumed is earlier in the cycle, with more upside remaining.
Check it: YIP (CoreLogic-backed) for growth and median, SQM Research for vacancy, Domain for days on market.
Both signals are public knowledge. The challenge is applying them across 8,417 Australian suburbs fortnightly, filtering the noise from thin markets where a handful of sales can distort the days-on-market figure beyond usability, and catching booms within weeks of starting rather than quarters.
That’s what the formula automates. Right now, 393 suburbs pass the hard filters — 35 under $400K, 149 under $600K, 204 under $800K — and receive fortnightly Strong Buy, Buy, Watch, or Pass labels based on their detection score.
What This Means for Evaluating Any Forecast
The next time you read a property forecast — from a bank, a research house, a buyer’s agent — three questions are worth asking:
How was accuracy measured?
A pooled correlation across all periods can look strong even when the model has no ability to rank suburbs within a single month. Ask for within-period discrimination results.
Were confidence ranges validated?
A model that claims 80% certainty should be right 80% of the time. If coverage isn't published, the confidence ranges are untested marketing.
What question is the model actually answering?
"Will this suburb grow?" in a rising market is almost always yes. The question that determines your relative return is: will this suburb outperform its peers in the same period?
We publish the full backtest methodology — the 78-suburb detection validation, the walk-forward tier discrimination results, and the forecaster failure analysis — on our proof page. No gating, no email required. The maths is there to check.
The core finding: prediction at suburb level with publicly available data delivers coin-flip accuracy. Detection — asking whether a suburb is already booming and how early you are — delivers 85.7% accuracy with a 20.2-point separation gap and zero false positives across 78 real outcomes.
Mean reversion is real too. Past outperformers tend to underperform going forward. The suburbs with the strongest 5-year momentum showed no cross-suburb ranking advantage within any given period. The data consistently rewards buying into early booms in affordable suburbs — not chasing past growth.
Join the Wishlist
We'll email you when BoomAU launches — starting with the budget range you care about.
Be first in line
- ✓Fortnightly Strong / Good / Fair / Weak signal labels per suburb
- ✓Filtered to your budget band
- ✓Built on a backtest of 12,360 postcode-months