Project Risk Lab
Interactive tools and research demos for evidence-based project decisions. Explore calibration, Monte Carlo simulation, reference class forecasting, and more.
Are You Well-Calibrated?
For each question, provide a 90% confidence interval: a range you are 90% sure contains the correct answer. If you are perfectly calibrated, 9 out of 10 answers should fall inside your ranges. Most people only capture 4–5 out of 10.
Monte Carlo Project Simulator
Define a base project cost and a set of risk events. Each risk has a probability of occurring and a cost impact range. The simulator runs multiple iterations to show the distribution of possible total project costs.
Risk Events
Reference Class Forecasting
Nobel laureate Daniel Kahneman advocates using the outside view: instead of relying on optimistic inside estimates, look at how similar projects actually performed. Select a project type to see historical cost overrun distributions from large-scale studies.
The Iron Law of Megaprojects
Based on the Oxford Global Projects database of thousands of projects worldwide:
Expected Value of Information
One of the most powerful (and most overlooked) tools in decision making. Before measuring anything, ask: what is the measurement worth? This tool calculates the Expected Value of Perfect Information (EVPI), the maximum you should spend to eliminate uncertainty before making a decision. It often reveals the measurement inversion: what we measure most has the least value, and what we ignore matters most.
How does this work?
You're deciding whether to proceed with a project. The decision hinges on an uncertain variable, such as expected revenue, NPV, or cost savings. If the true value falls below your threshold, proceeding would be a mistake. EVPI quantifies the maximum you should pay to resolve that uncertainty before committing.
Noise Audit
Even perfectly unbiased cost estimates produce systematic cost overruns. When organisations pick projects based on estimated value for money, they inadvertently favour projects whose costs were underestimated by luck. This "winner's curse" has nothing to do with bias or gaming. This tool quantifies how much of the observed overrun for a given project type could be explained by estimation noise alone, without any bias, optimism, or strategic misrepresentation.
How does the budget constraint create overruns?
The model’s key mechanism is the budget constraint acting as a selection filter. Organisations cannot fund every project, so they rank proposals by estimated benefit-cost ratio (BCR) and fund the top fraction (the budget share α). Among funded projects, costs systematically exceed estimates because the project was selected partly because noise made its costs look lower than they really were. This is the winner’s curse applied to project portfolios.
Two parameters govern the effect: σ (how noisy the estimates are) and α (how selective the budget is). Tighter budgets (lower α) force selection from the extreme tail of the estimated BCR distribution, amplifying the overrun. When α = 1 (everything funded), there is no selection and noise averages out.
Selection-Adjusted RCF
Standard Reference Class Forecasting applies the full observed overrun as an uplift (a percentage added to the estimate to account for historical patterns of cost growth). But part of that overrun is a selection effect (winner’s curse): projects were funded partly because noise made them look cheaper. Since this selection effect will recur for the new project, correcting for it means overcorrecting. This tool strips the selection component, producing an uplift based on genuine bias only.
Why does naive RCF overcorrect?
When organisations rank projects by estimated benefit-cost ratio (BCR) and fund the top fraction α, projects with accidentally underestimated costs are more likely to be selected. The observed overrun in the reference class therefore reflects both genuine estimation bias and this statistical selection effect (a winner’s curse). But the new project being appraised will go through the same selection filter, so the selection-driven overrun will happen again regardless of any uplift. Applying the full observed overrun as an RCF adjustment therefore double-counts the selection component.
This tool decomposes the observed overrun into a selection component (teal) and a residual bias component (gold) using a closed-form expression for the expected cost ratio conditional on BCR-based selection, then constructs a corrected forecast curve that adjusts only for the bias portion.
- The correction only matters when the RCF uplift is applied before the selection decision (i.e., during appraisal). If applied after funding, naive RCF is appropriate.
- The decomposition into selection and bias components is exact for the mean. The percentile-level correction is a first-order approximation that applies the mean-level bias fraction to the overrun (cost) or shortfall (benefit) portion of each percentile, leaving under-budget or above-target outcomes unchanged.
Conditional Reference Class Forecasting
Standard reference class forecasting applies the outside view at the start of a project. But what happens as the project progresses and you observe actual performance? Conditional RCF updates the forecast by combining the reference class distribution with real progress data, narrowing uncertainty as evidence accumulates.
What makes it conditional?
All reference classes are conditional. When you pick "IT cost overruns" as your reference class, you've already conditioned on project type. The term Conditional Reference Class Forecasting adds a further condition: how far along the project is and what actual performance looks like so far. At 0% progress, you have only the unconditional distribution. At 50% progress with an observed cost ratio of 1.15, you combine that evidence with the reference class to produce a tighter, more informed forecast.
Beta Distribution
How uncertain are you about a proportion, like the probability of project cancellation, the share of requirements that change, or the fraction of stakeholders who approve? The beta distribution is the Bayesian tool for exactly this question. Start with a prior belief, add observed data, and watch the uncertainty narrow.
How do α and β work?
The parameters α (alpha) and β (beta) shape the distribution. Think of them as pseudo-counts: α = prior "hits" + 1 and β = prior "misses" + 1. Start with α=1, β=1 for a flat (uninformative) prior. Larger values indicate stronger prior belief, meaning a sharper peak and less spread. When you observe new data (hits and misses), the posterior is simply Beta(α + hits, β + misses).
Prior Belief
Observations
Technology Regret Analysis
Technology costs tend to decline while benefits grow over time. Adopt too early and you overpay; too late and you forfeit years of benefits. Technology regret is the NPV lost by investing at a non-optimal time. This tool computes the net present value of investing in each possible year and identifies the optimal adoption window.
What drives the optimal timing?
Two forces compete: cost decline rewards waiting (technology gets cheaper), while the discount rate and finite benefit window penalize delay (future benefits are worth less, and the planning horizon shrinks). The model discounts all cash flows to year 0, so you can compare adopt-now vs. adopt-later on equal footing. The optimal year balances the savings from cheaper technology against the benefits forgone by waiting.
Risk Matrix Roulette
Think risk matrices give you a clear picture? Play three rounds that expose fundamental flaws in the most widely used risk assessment tool. Based on research by Cox (2008), Mandel (2021), and Hubbard, Budzier & Leed (2025).
Why challenge risk matrices?
Risk matrices are ubiquitous in project management, but research shows they suffer from at least three fatal flaws: (1) verbal probability labels mean vastly different things to different people, (2) colour-coded cells hide enormous differences in expected loss, and (3) arbitrary category boundaries can reverse the ranking of risks. This interactive challenge lets you experience all three.
Parade of Trades
Five trades. Thirty-five rooms. One die. Can you beat the plan? In construction, trades work sequentially through rooms, but variability in weekly output creates cascading delays. Roll the dice and watch the parade unfold.
How does this work?
Five trades (Civil, Mechanical, E&I, Fit-out, Commissioning) must each process 35 rooms in sequence. Each trade rolls a die each week to determine capacity, but can only process rooms the previous trade has already completed. The deterministic plan, using the average roll of 3.5 rooms/week, expects completion in 14 weeks. You will play three rounds with dice of identical averages but different variability to see how uncertainty amplifies through the system.
This demonstrates a general principle: when tasks must happen in sequence and each step has variable output, the total project duration is always longer than the sum of averages. Planning with averages guarantees you will be late.
Probability Word Explorer
What does “likely” mean? What about “probable”? In a survey of 123 respondents, people assigned wildly different numeric probabilities to common probability words. Explore all 17 words and see how your own interpretations compare.
About the data
This data comes from a survey conducted by Wade Fagen-Ulmschneider (Teaching Professor, Computer Science, University of Illinois at Urbana-Champaign). Each of the 123 internet survey respondents was asked to assign a numeric probability (0–100%) to 17 common probability words. The box-and-whisker chart below shows the distribution of responses for each word: the box spans the interquartile range (Q1–Q3, the middle 50% of responses), the vertical line marks the median, and the whiskers extend to the minimum and maximum values observed. The study builds on a long tradition of research into verbal probability, notably Sherman Kent’s 1964 “Words of Estimative Probability” for the CIA.
Go Deeper
Several of these tools are based on methods from How to Measure Anything in Project Management. The book provides the full framework for evidence-based project decisions.