There’s a chart your mutual fund app shows you: “Investors who stay invested for 10 years earn 14% annually on average.” That chart is real. The advice is correct. And almost nobody follows it. Not because people are stupid — because knowing something and doing it are completely different problems. The entire wealth management industry has spent fifty years solving the wrong one. They’ve been obsessed with what to invest in. Nobody has seriously asked why people stop investing. India’s SIP stoppage ratio hit 109% in January 2025 — more SIPs cancelled than opened, for the first time in history. Causal AI is starting to answer why.
The Ice Cream Problem
In 1970s New York, city officials noticed something alarming: on days when ice cream sales spiked, so did the murder rate. The correlation was strong, consistent, and reproducible across years of data.
The actual explanation is summer. Hot weather causes people to buy ice cream. Hot weather also drives people outside, increases social friction, and causes more violence. Ice cream and murders share a common cause. Neither caused the other.
This is so obvious in retrospect that it’s almost funny. But the same mistake — confusing correlation with causation — is embedded in virtually every financial model ever built, every robo-advisor ever launched, and every “AI-powered” investment app you’ve ever used.[1]
Marcos López de Prado at Cornell and ADIA Lab showed with 10,000 Monte Carlo simulations that correlation-based factor models frequently produce portfolios that buy what they should sell and sell what they should buy. He called the result a “factor mirage.”
When a wealth platform builds its recommendation engine, it looks at historical data and asks: what patterns predict good outcomes? “Users who rebalance quarterly outperform.” “Investors who pause SIPs during corrections underperform.” All of this is correlation. The data is real. But correlation is not the same as causation, and the difference matters enormously when you’re trying to do something about it.
Correlation ≠ Causation: The Ice Cream & Murder Puzzle
Here’s a concrete example. Suppose you observe that investors who check their portfolio frequently have better long-term returns. Correlation is clear. So you design a feature to make people check more. Does this work? Probably not. People who are financially committed both check their portfolio more and have better returns. The checking didn’t cause the returns. Both were caused by the same upstream thing: genuine financial engagement.
Confounders & Colliders
Before going further, two concepts that explain exactly how correlations mislead you.
A confounder is a hidden third variable that causes both things you’re observing. Summer causes both ice cream and murders. Financial commitment causes both app usage and SIP continuation. If you don’t measure the confounder, you’ll incorrectly conclude one thing caused the other.
A collider is trickier. It’s a variable that is caused by two things you’re measuring. And when you try to control for it, you create a fake relationship between those two things that wasn’t there before.[2]
The classic example: shoe size and reading ability are unrelated. But both are caused by age. If you filter to only look at children of the same age, shoe size and reading ability suddenly look correlated — because you’ve controlled for the collider (age) and manufactured a relationship that doesn’t exist.
In finance, this happens constantly. Adding more variables to your model to make it look sophisticated often makes it worse — not because the variables are irrelevant, but because some are colliders, and including them corrupts everything else.
López de Prado’s 10,000 simulations. With 10,000 Monte Carlo trials, he showed that correlation-based factor models — the kind used by every major asset manager — frequently produce portfolios that buy what they should sell and sell what they should buy. He called the result a “factor mirage” — a pattern that looks real in backtests but evaporates with real money.
Confounder vs Collider: Two Ways Correlations Lie
Pearl’s Ladder
Judea Pearl — who won the Turing Award in 2011 for his work on causal inference — describes three levels of causal reasoning. They map almost perfectly onto the gap between what current wealth apps do and what’s actually possible.
Level 1: Observation. “What happened?” Your portfolio dropped 8%. Your SIP continuation rate is 70%. This is what every platform tells you.
Level 2: Intervention. “What happens if I do X?” If we send a goal reminder, what’s the probability they keep their SIP? This requires a causal model — you can’t answer intervention questions with pure correlation.
Level 3: Counterfactual. “What would have happened if?” You stopped your SIP in February 2022. If you had stayed invested, your portfolio would be ₹2.3 lakhs larger. This level of reasoning is only possible with a structural causal model.[3]
Pearl’s “Ladder of Causation” is formalized in his 2009 book Causality and popularized in The Book of Why (2018). The three levels correspond to increasing levels of causal reasoning that no amount of data alone can bridge.
Every current platform operates primarily on Level 1. Some gesture toward Level 2. Nobody has seriously built Level 3. But Level 3 is the one that actually changes behavior — because people don’t respond to abstract advice. “Stay invested for the long term” bounces off. But “if you had stayed invested through the last correction, you’d have ₹2.3 lakhs more right now” — that gets through.
Pearl’s Ladder of Causation: Three Levels
The counterfactual is a mirror. And mirrors change people in a way that charts don’t.
The Causal Graph
Causal AI starts from a different question. Instead of “what patterns exist in this data?”, it asks “what actually causes what?” The machinery involves a causal graph — a diagram that explicitly maps which variables cause which other variables. Not just which ones correlate.
Here’s a causal graph for why an investor stops their SIP. A market drop triggers loss aversion, which leads to “I’m losing money, this was a mistake,” which forms a cancellation intent, and then — crucially — there’s a 21-day window before the SIP is actually cancelled.
A correlation-based system sees: “market down → SIP cancellations increase.” A causal system discovers: there’s a 21-day intervention window. And the right intervention isn’t market reassurance — it’s goal re-anchoring. Reminding the investor of the specific thing they’re investing for. When the goal becomes vivid again, short-term noise becomes irrelevant.
SIP Cancellation Causal Graph
The intervention type matters, not the intensity. The industry sends market reassurance: “Markets are volatile. Stay the course.” Research on loss aversion shows this has near-zero effect. People in the grip of loss aversion are moved by something more primal — the vivid reality of what they’re trying to build. Their children’s education. Their freedom from financial anxiety. Reconnecting someone to their goal changes behavior in a way that market reassurance doesn’t.
The Engagement Trap
Here’s the specific mistake that most consumer fintech companies make. Engagement — how often users open the app, check their portfolio, read notifications — correlates with retention. Users who engage more tend to stay invested longer. So product teams optimize for engagement. More notifications. Gamification. Streaks. Confetti animations.
This doesn’t work. Engagement and retention share a common cause: financial commitment. People who are genuinely committed both engage more and stay invested longer. The engagement didn’t cause the retention.
In causal terms: engagement is a collider. Both commitment and retention cause engagement. When you control for engagement — by trying to maximize it — you open a spurious relationship. Your model thinks the nudges are working, but you’re really just measuring the committed users and calling the difference “impact.”[4]
The correct intervention is upstream. Not “get the user to engage more” but “increase the user’s financial commitment” — make the goal more concrete, vivid, and emotionally real. Engagement follows naturally.
The Engagement Trap: Collider Bias in Practice
The users who actually need help are the ones who stop engaging. The quiet ones. The ones who don’t open the app for three weeks, then cancel their SIP on a Tuesday morning. Those users never received an intervention, because the engagement-optimization system had already classified them as “low-engagement” and deprioritized them.
The Counterfactual Mirror
What if, instead of a portfolio dashboard, your investment app showed you this: “You’ve started 3 SIPs in the last 4 years. You’ve stopped all 3. Here’s why — not what you might think.”
Each cancellation has a different root cause. One was pure loss aversion during a market dip. One was a calendar mismatch between SIP debit and salary credit dates. One was goal ambiguity — “general savings” was too vague to survive a competing priority. The total cost across all three: ₹74,400 in compounding.
Nobody has ever shown a user this. Because nobody has ever built the causal model required to produce it. This isn’t a feature. It’s a different kind of relationship between a product and a user — one built on accurate diagnosis rather than generic advice.
The Counterfactual Mirror: What If You Had Stayed?
Why this is hard. Causal models require intellectual humility (you must commit to testable assumptions), longitudinal data (the same users observed over years), domain knowledge (you can’t learn causal direction from data alone), and new tooling (DoWhy, EconML, tigramite). All barriers are real. They’re also exactly why the competitive moat is deep.
Causal AI is not magic. It doesn’t predict the market. It doesn’t guarantee returns. What it does is give you an honest accounting of what’s actually happening — in your portfolio, in your behavior, in the interventions that do and don’t work. That’s the difference between knowing what you should do and actually doing it. Between advice and intervention. Between a chart and a mirror.
Based on the work of Judea Pearl (Causality, 2009), Marcos López de Prado et al. (Causal Factor Analysis, ADIA Lab, 2025), Susan Athey & Stefan Wager (Causal Forests, 2018), and Victor Chernozhukov et al. (Double Machine Learning, 2018). Market data from AMFI monthly SIP statistics.