Why single-line projections mislead

Every standard financial plan contains a chart that looks roughly the same: a single line ascending from left to right, inflected upward by the assumed rate of return, terminating at some target wealth figure in retirement. It is clean, legible, and almost entirely misleading. The line communicates a false sense of precision, presenting one deterministic trajectory when in reality the space of plausible outcomes spans an enormous range.

The root of the problem is the arithmetic versus geometric return distinction — one of the most consequential and consistently misunderstood concepts in quantitative finance. Arithmetic average returns and geometric (compound) returns are not the same thing. Volatility drives them apart, and the gap widens with time. The relationship is approximated as: geometric return ≈ arithmetic return − (volatility² / 2). A portfolio with an arithmetic average return of 7% and annual volatility of 15% has an expected geometric return of roughly 5.9%. Over 30 years, the compounding difference between those two numbers is enormous.

"A portfolio that returns +50% then −33% has an arithmetic average of +8.5% per year, but a geometric return of exactly 0%. The terminal value is identical to the starting value."

This is not an edge case. It is the rule. Whenever an advisor or planning tool inputs a single "expected return" figure into a deterministic projection, they are implicitly assuming that return is achieved smoothly, every year, without volatility. The real world delivers returns in a lumpy, path-dependent sequence that may have the same average but a radically different terminal value. Single-line projections do not capture this. They do not capture the probability of falling 40% in year three of retirement. They do not show the client what their portfolio looks like at the 10th percentile of outcomes. They present a single expected-case number that is, by construction, the median of nothing — it is simply an arithmetic extrapolation that conflates the mean of a distribution with a likely outcome. Monte Carlo simulation is the correction to this structural flaw.

What Monte Carlo simulation actually does

The term "Monte Carlo" refers to a broad class of computational methods that use repeated random sampling to obtain numerical results. In the context of portfolio planning, the methodology is conceptually straightforward: define a return distribution for each asset class in the portfolio, specify their correlations, draw random return sequences for each period across a chosen time horizon, and compute the resulting portfolio value. Repeat this process thousands of times — typically 10,000 paths is standard — and collect the distribution of terminal outcomes.

The return distributions can be constructed in two principal ways. The first is parametric: assume that returns follow a known distribution (most commonly normal, or Gaussian) characterised by a mean and standard deviation estimated from historical data or set as forward-looking assumptions. The second is historical bootstrapping: draw actual historical return sequences at random, with replacement, preserving the correlation structure across asset classes in each sampled period. Each approach has trade-offs. Parametric simulation is clean and controllable but depends critically on distributional assumptions — a point I will return to in the limitations section. Historical bootstrapping preserves empirical fat tails and cross-asset relationships but is constrained by the length of the historical record and may not capture forward-looking regime changes.

The output of 10,000 simulated paths is visualised as a fan chart: a band of possible portfolio trajectories that narrows near inception (all paths start from the same point) and widens with time as the paths diverge. The fan is typically shaded to show percentile bands — the 10th through 90th percentile corridor, the median line, and sometimes the 25th and 75th percentile inner band. This visual representation communicates something that a single line fundamentally cannot: the outcome is genuinely uncertain, the uncertainty compounds over time, and planning should account for the full distribution rather than the single expected value. The covariance structure underlying these simulations — the correlations and volatilities that determine how assets co-move — benefits enormously from rigorous estimation techniques such as Ledoit-Wolf shrinkage, which I discuss in detail in a companion article.

Sequence-of-returns risk: why order matters as much as average

Of all the concepts that Monte Carlo simulation makes legible, sequence-of-returns risk is perhaps the most important for investors approaching or in retirement. The arithmetic seems counterintuitive at first: how can two investors with identical 20-year average annual returns end up with dramatically different terminal wealth? The answer is that in the presence of ongoing contributions or withdrawals, the order in which returns are received materially affects the outcome in a way that the average return does not capture.

The asymmetry is structural. During the accumulation phase — when an investor is adding contributions to a portfolio each month — poor early returns are partially mitigated by the ability to purchase additional units at depressed prices (this is the mechanism behind dollar-cost averaging). A severe drawdown in year two of a 30-year accumulation horizon is painful but partially self-correcting. The distribution phase reverses this entirely. Once an investor begins withdrawing from a portfolio — in retirement, or during any structured drawdown — poor early returns force the sale of units at depressed prices to fund spending. Those units are permanently gone. They cannot participate in the subsequent recovery. The portfolio is permanently impaired.

Worked Example — Sequence of Returns

Investor A retires with $1,000,000 and withdraws $50,000 per year. Returns over the first four years: −25%, −15%, +30%, +25%. Four-year arithmetic average: +3.75%.

Investor B retires with the same $1,000,000 and withdraws the same $50,000 per year. Returns in reverse order: +25%, +30%, −15%, −25%. Four-year arithmetic average: also +3.75%.

Investor A ends year four with approximately $612,000. Investor B ends year four with approximately $786,000 — a difference of $174,000, entirely attributable to the sequence in which identical average returns arrived. Extend this asymmetry across a 20-to-30-year retirement and the outcomes diverge catastrophically at the lower end of the distribution.

Monte Carlo simulation captures this risk precisely because it generates thousands of distinct return sequences, including sequences where severe drawdowns cluster early in retirement. A deterministic projection using the average return will always show the same optimistic terminal value. The simulation shows you the distribution of outcomes, and critically, it shows you how many of those thousands of paths ended in portfolio depletion before the planned end date. That number — the probability of ruin — is one of the most consequential pieces of information available to a retiree, and it is invisible in any single-line projection. This is also why tail risk measurement is a natural complement to Monte Carlo analysis: where the simulation tells you the frequency of bad outcomes, CVaR tells you the severity.

How to read a Monte Carlo output

The output of a properly constructed Monte Carlo simulation should be read as a probability distribution, not as a prediction. This distinction is essential. When I run 10,000 simulations for a client's portfolio and report that 78 of those paths ended in ruin, I am not predicting a 22% chance of ruin with any particular precision. I am saying that under the specified return and volatility assumptions, and under the specified spending rate, 22% of plausible historical return sequences would have caused portfolio depletion. This is qualitatively different from a deterministic forecast, and it should be used differently.

The three numbers I focus on with clients are the 10th percentile planning floor, the 50th percentile central case, and the 90th percentile optimistic scenario. The 10th percentile is the most important of the three. It represents the portfolio value at which only 10% of simulated paths performed worse — it is, in practical terms, the floor against which one should stress-test spending plans. A retirement that remains solvent at the 10th percentile is reasonably robust to adverse sequences. A retirement that runs into difficulty at the 30th percentile is fragile.

The "success rate" or "goal probability" figure that Monte Carlo tools report — typically stated as "your plan has a 78% success rate" — means precisely that 78 of 100 simulated paths achieved the stated goal (most commonly, maintaining a positive portfolio balance through a specified retirement age). This is actionable information. If the success rate is 95%, the plan is robust. If it is 60%, the plan requires adjustment — either reduced spending, increased savings, extended working years, or some combination. The simulation does not tell you which lever to pull; it tells you how hard you need to pull. Clients can use the Risk Assessment tool to calibrate their own risk tolerance before interpreting these percentile outcomes in a personal context.

One practical point: the success rate figure is sensitive to the assumed time horizon and spending rate. A 4% withdrawal rate on a 30-year horizon produces a materially lower success rate than a 3.5% rate on the same horizon, and the relationship is nonlinear in the tail. Small reductions in the withdrawal rate produce disproportionate improvements in the probability of success because they reduce the frequency of the worst-case compounding failures. This insight — that modest spending flexibility has outsized impact on tail outcomes — is one of the most practically valuable results that Monte Carlo analysis produces.

Practical inputs and their impact

The quality of a Monte Carlo analysis is entirely determined by the quality of its inputs. Garbage in, garbage out — and in portfolio simulation, the garbage is often invisible because the output looks numerically precise regardless of how poorly the inputs were specified.

The most consequential input is the return assumption. I use forward-looking return estimates rather than simply extrapolating historical averages. The reason is straightforward: current valuations, yield environments, and structural conditions in markets embed meaningful information about prospective returns that the historical average ignores. Using the historical equity premium mechanically in an environment of elevated valuations and compressed credit spreads overstates the expected return and systematically understates the probability of poor outcomes. Forward-looking inputs require judgment, but they produce more honest projections.

Volatility and correlation inputs are the second most important determinant of simulation quality. I estimate the covariance matrix using Ledoit-Wolf shrinkage, a regularisation technique that improves on the sample covariance matrix by shrinking it toward a structured target. The sample covariance matrix — computed directly from historical returns — is notoriously noisy, particularly when the number of assets is large relative to the number of observations. Ledoit-Wolf shrinkage produces a better-conditioned estimate that reduces the impact of estimation error on simulation outcomes. The improvement in covariance estimation translates directly to more reliable tail risk estimates in the Monte Carlo output.

Spending rules and withdrawal rates interact with the simulation in important and nonlinear ways. A fixed dollar withdrawal (e.g., $80,000 per year regardless of portfolio value) produces different and generally worse tail outcomes than a variable spending rule (e.g., withdrawing a fixed percentage of current portfolio value, or using a floor-and-ceiling guardrail approach). Variable spending rules effectively build a form of automatic adjustment into the plan — when the portfolio underperforms, spending adjusts downward, reducing the probability of ruin at the cost of spending variability. The choice of spending rule is as important as the return assumption in determining success rates.

Finally, time horizon sensitivity deserves explicit attention. Extending the planning horizon by five years does not increase risk linearly — it increases the probability of encountering a sustained adverse sequence and amplifies the compounding effect of early drawdowns. This is particularly relevant for clients who may live significantly longer than median life expectancy. Planning to the median is, by construction, planning to fail half the time.

What Monte Carlo tells you that a financial plan doesn't

A traditional financial plan — even a sophisticated one — is a deterministic document. It projects a single path forward based on assumed returns, spending, inflation, and timeline. Monte Carlo simulation replaces that single path with a distribution, and in doing so, it surfaces four categories of insight that the deterministic plan cannot produce.

First, it quantifies tail scenarios and ruin probability. The probability that a portfolio is depleted before the end of the planning horizon is not just a number — it is a calibration of how much buffer the plan contains. A plan showing 95% success has very different characteristics than one showing 75% success, even if both show the same median terminal value. The difference is entirely in the distribution of outcomes below the median, and it is precisely what determines whether a client sleeps well or not.

Second, Monte Carlo reveals the optimal spending rate as a function of desired confidence level. Rather than applying a rule-of-thumb withdrawal rate (the "4% rule" being the most cited, though its applicability in current market conditions warrants scrutiny), the simulation allows explicit calibration: what withdrawal rate corresponds to a 90% success rate, a 95% rate, or a 99% rate? The answer depends on the specific asset allocation, return assumptions, and time horizon — it is not a universal constant. The Asset Lens tool is designed to help run this analysis across different portfolio compositions.

Third, Monte Carlo makes explicit the value of flexibility. A client who is willing to reduce spending by 10% in response to a poor sequence of returns dramatically improves their probability of success. The simulation quantifies this: spending flexibility is a form of risk management, and it can be traded directly against asset allocation risk. A more conservative portfolio paired with flexible spending often outperforms an aggressive portfolio with rigid spending at the tail of the distribution.

Fourth, Monte Carlo provides the analytical foundation for comparing the impact of different risk management strategies — such as incorporating alternative allocations, adjusting asset class weights, or evaluating the benefit of structured products — on the full distribution of outcomes rather than just the expected return. This connects naturally to the Black-Litterman framework for expressing return views within a disciplined optimisation, where the optimised portfolio is subsequently stress-tested through simulation.

Limitations: model risk and what Monte Carlo cannot capture

Monte Carlo simulation is a planning tool, not a forecast. This distinction is not merely semantic — it is essential to using the methodology responsibly. Every simulation result is conditional on its assumptions, and those assumptions introduce model risk that can be as consequential as the market risk being modelled.

The most significant limitation is the distributional assumption. Most Monte Carlo implementations assume that returns are drawn from a normal (Gaussian) distribution. Empirical asset returns are not normal. They exhibit fat tails — extreme events occur far more frequently than a normal distribution predicts. The 2008 financial crisis, the March 2020 drawdown, and the 1987 crash all involved return realisations that a normal distribution assigns near-zero probability. A simulation built on normality systematically underestimates the frequency and severity of tail events. Historical bootstrapping preserves some of this tail behaviour, but it is constrained by the historical record, which may not contain analogues for future tail events.

The second limitation is parameter estimation error. The return and volatility assumptions that feed the simulation are themselves estimates, subject to substantial uncertainty. A 1% error in the assumed real return has enormous consequences for 30-year projections. Users of Monte Carlo outputs should treat the success rate as an order-of-magnitude indicator — the difference between 92% and 94% is not meaningful; the difference between 70% and 90% is. Precision in the output does not reflect precision in the inputs, and it is a persistent failure of financial planning software to display results to decimal places that imply far more confidence than the methodology supports.

Third, Monte Carlo as typically implemented does not capture structural breaks and regime changes. The simulation draws from a single stationary distribution across the entire horizon. In reality, the economy and markets undergo regime shifts — transitions between high-growth, low-growth, inflationary, and deflationary regimes — that alter the return-generating process. A simulation calibrated on post-1990 U.S. data, for instance, will not produce return sequences that look like the 1970s stagflation environment or Japan's post-1990 deflation. Regime-switching models address this partially, but they introduce additional layers of parameter uncertainty. At the advisory level, the appropriate response is scenario analysis alongside simulation — complementing the probabilistic output with deterministic stress tests of specific adverse regimes.

None of these limitations diminish the value of Monte Carlo simulation relative to deterministic projections. They simply define the tool's epistemic boundaries. A simulation that correctly models the distribution of outcomes under stated assumptions, and that is transparently paired with scenario analysis for events outside those assumptions, is the most rigorous planning framework available outside a dedicated quantitative research desk. Used with appropriate humility about its assumptions, it transforms financial planning from a false precision exercise into an honest accounting of risk and probability.