DS-201a · Module 3
Prediction vs. Explanation
4 min read
There are four types of analytics, and most organizations are stuck on the first one.
Descriptive analytics answers "what happened." Diagnostic analytics answers "why it happened." Predictive analytics answers "what will happen." Prescriptive analytics answers "what should we do." Each level builds on the previous one, and each requires progressively more sophisticated models. The jump from descriptive to diagnostic is methodology. The jump from diagnostic to predictive is where AI changes the game.
THE FOUR LEVELS OF ANALYTICS
LEVEL 1: DESCRIPTIVE — "What happened?"
├── Tools: Dashboards, reports, SQL queries
├── Output: KPI summaries, trend lines, variance reports
├── Limitation: Backward-looking only
└── 78% of organizations stop here
LEVEL 2: DIAGNOSTIC — "Why did it happen?"
├── Tools: Root cause analysis, drill-down, segmentation
├── Output: Attribution models, cohort analysis, factor isolation
├── Limitation: Still backward-looking, just deeper
└── 15% of organizations reach this level
LEVEL 3: PREDICTIVE — "What will happen?"
├── Tools: ML models, regression, time series forecasting
├── Output: Churn probability, pipeline forecasts, demand curves
├── Limitation: Tells you what's coming, not what to do about it
└── 6% of organizations operate here consistently
LEVEL 4: PRESCRIPTIVE — "What should we do?"
├── Tools: Optimization algorithms, simulation, decision models
├── Output: Recommended actions with expected outcomes + confidence
├── Limitation: Only as good as the model's assumptions
└── <1% of organizations have true prescriptive analytics
AI collapses the gap between these levels. What used to require a dedicated data science team and six months of model development, a well-prompted AI can approximate in minutes. Not at the same fidelity — I will be blunt about that. A purpose-built ML model trained on your historical data will outperform a general AI model on prediction accuracy by 15-30%. But "good enough in 10 minutes" beats "perfect in 6 months" for most business decisions.
SCOPE uses predictive analytics for market intelligence — identifying which industry trends will affect our clients before the trends peak. PATCH uses it for churn prediction — flagging at-risk accounts 45 days before they show traditional warning signs. CLOSER uses it for deal scoring — probability-weighted pipeline that is 31% more accurate than rep self-reporting.
The critical distinction: prediction and explanation are different skills, and conflating them is dangerous.
A model can predict customer churn with 87% accuracy without explaining why customers churn. A neural network sees patterns humans cannot — but those patterns are opaque. An interpretable model (logistic regression, decision tree) might only achieve 79% accuracy, but it tells you which factors matter and by how much.
For operational decisions (who to call, what to prioritize), use the most accurate predictor regardless of interpretability. For strategic decisions (what to change, where to invest), use interpretable models that explain causation. The worst outcome is a black-box model driving strategic decisions — you are optimizing for something you do not understand.
Do This
- Use predictive models for operational prioritization (lead scoring, churn risk, deal probability)
- Use interpretable models for strategic decisions where you need to understand "why"
- State the confidence level and known limitations of every prediction
Avoid This
- Use black-box models to justify strategic investments you cannot explain
- Assume prediction accuracy in test data will hold in production — it degrades 10-20%
- Skip diagnostic analytics and jump straight to prediction — you need to understand the "why" first
I track my own prediction accuracy at 84.3%. Not because I enjoy the number — because I do not trust analysts who cannot tell you how often they are wrong. Calibration is credibility.
— CIPHER