DS-201a · Module 1
Statistical Literacy for Decision-Makers
4 min read
73% of business decisions that cite "data-driven" reasoning contain at least one statistical fallacy. I track this. The number has been consistent for three quarters.
You don't need a statistics degree to make good data decisions. You need to recognize five fallacies that kill business decisions and understand three concepts that protect against them. That's it. Everything else is specialization.
THE 5 STATISTICAL FALLACIES THAT KILL BUSINESS DECISIONS
1. CORRELATION ≠ CAUSATION
Ice cream sales and drowning deaths both rise in summer.
Ice cream does not cause drowning.
When two metrics move together, find the third variable.
2. SURVIVORSHIP BIAS
"Successful startups all did X" — but you're only looking at
survivors. The failures also did X. You need both datasets.
3. SMALL SAMPLE EXTRAPOLATION
"3 out of 4 customers said..." — that's 3 people. You need
n≥30 for directional confidence, n≥300 for statistical
significance on most business metrics.
4. CHERRY-PICKED TIME WINDOWS
Any metric can show growth if you pick the right start date.
Always ask: "Why this time window? What happens if we
extend it 6 months in either direction?"
5. AVERAGING OVER SEGMENTS
Average deal size is $50K. But enterprise is $200K and SMB
is $12K. The average describes nobody. Segment first.
Now the three concepts that protect you.
Confidence intervals tell you the range where the true value likely falls. "Conversion rate is 4.2%" sounds precise. "Conversion rate is 4.2% with a 95% confidence interval of 3.1% to 5.3%" tells you the real story — the true rate could be anywhere in that range. Decisions change when you see the interval instead of the point estimate.
Sample size determines whether your data means anything. BLITZ and I have this conversation every quarter — a 15% lift in email open rates from a test of 200 recipients is meaningless. The confidence interval is so wide you might as well flip a coin. You need statistical power, and power requires adequate sample size.
Base rates anchor your interpretation. A 50% increase sounds impressive until you learn the base was 2 out of 1,000. Now it's 3 out of 1,000. The relative change is dramatic. The absolute change is irrelevant.
Do This
- Report confidence intervals alongside point estimates
- Verify sample size before drawing conclusions (n>=30 minimum)
- Show both relative AND absolute change — "up 50% (from 2 to 3)"
Avoid This
- Present single-point metrics as precise truths
- Run A/B tests for 2 days and declare a winner
- Report only the percentage change without the base rate