1 in
Prevalence: 0.01%
1% 20% 40% 60% 80% 100%
0.1% 20% 40% 60% 80% 90%
Adjust parameters and click "Generate Results"
Truth Failed Test (Flagged) Passed Test (Not Flagged) Total

📊 Results Summary

How This Works

This tool corrects for base rate neglect (also called base rate bias or the base rate fallacy) - a common cognitive bias where people ignore how the prevalence of a problem affects test outcomes.

The simulator applies Bayes' theorem to show you four outcome categories instead of just one accuracy number. It displays results in frequency format (actual counts of people) rather than percentages, which helps our intuitions work better when reasoning about probabilities.

Even highly accurate tests can produce more false positives than true positives when screening for rare conditions. This has profound implications for mass screening programs in health, security, and education.

📖 Want to learn more? Read the full explanation of why rare-event screening is so dangerous and how this tool fits into the bigger picture: Rarity Roulette: Making the Math of Mass Screenings Visible

Real-World Examples

Mass screenings for low-prevalence problems appear across many domains:

Note: The mammography and PSA examples use parameters extrapolated from the Harding Center's long-term screening data to match their published false-alarm rates. The Harding fact boxes provide more comprehensive outcomes (e.g., mortality, overdiagnosis, quality-of-life effects), whereas this tool focuses solely on test-classification accuracy. The purpose is illustrative, not a substitute for full cohort data.

Important Considerations

These outcome estimates assume the stated accuracy rates are correct and precisely known. In practice, several factors may distort these estimates:

Why accuracy may be inflated:

  • Tests often perform better in controlled lab settings than in real-world field conditions
  • Accuracy parameters themselves carry uncertainty (see Gelman & Carpenter, 2020)
  • Researchers and tool developers have financial and professional incentives to report favorable results

Why error rates may be misleading:

  • False negatives may be underestimated when dedicated adversaries actively game the system (e.g., espionage, fraud)
  • Detection thresholds for continuous biomarkers are often set arbitrarily
  • Test performance may vary substantially across demographic subgroups

Hidden costs of false positives:

  • Secondary screenings may cause unintended harm (e.g., elevated mortality risks from unnecessary radiation/other treatment, chronic pain from unnecessary biopsy or mastectomy, psychological distress, and job loss)
  • Those who benefit financially from follow-up testing often set screening guidelines
  • Individuals flagged incorrectly bear costs while having no actual condition

What this tool doesn't capture:

  • This calculates hypothetical estimates along only one causal pathway — the classification pathway
  • It does not compute net effects accounting for test classification, strategy, information, and resource reallocation