“Why did it happen? That’s the ultimate question human beings ask, and understanding causation helps us answer it.”

1. The Misinterpretation of Data: Correlation Is Not Causation

For years, people have clung to the idea that correlation does not imply causation, leaving a gap in how we understand events and outcomes. This mindset led many statisticians to dismiss causative theories as scientifically invalid. Karl Pearson, a prominent statistician, famously reduced science to "pure data," arguing causation couldn’t be proven and was thus unscientific.

The problem is that rejecting causation outright ignores hidden connections. Consider Pearson’s favorite around chocolate and Nobel laureates: wealthier nations consume more chocolate and also fund research, leading to more Nobel wins. The dismissal of causation undervalues these underlying factors.

Sewall Wright showed in 1912 that causation could be mathematically represented. Using guinea pigs to study inherited coat patterns, he created path diagrams to demonstrate how heredity and development contribute. His work, though buried for decades, signaled the start of a deeper understanding of causation.

Examples

  • Karl Pearson’s rejection of causation and his chocolate/Nobel argument.
  • Sewall Wright’s use of path diagrams to prove genetic inheritance in guinea pigs.
  • Wealth as a causative link between chocolate consumption and Nobel Prize production.

2. Data Without Causal Analysis Can Be Misleading

Data alone cannot unveil how or why certain outcomes happen. For example, early analysis of the smallpox vaccine wrongly suggested it caused more deaths than the disease itself because the data didn’t account for what would happen without the vaccine.

Improper understanding leads to strange conclusions. For instance, the observed connection between children’s shoe sizes and reading ability stems from age, not shoe size influencing literacy. Without addressing this common cause, conclusions become baffling.

Recognizing connections beyond surface data is vital. For this reason, the book introduces the Ladder of Causation, a framework for exploring causality instead of passively relying on data alone.

Examples

  • Early vaccine data appearing to blame immunization for fatalities.
  • Shoe size and literacy being linked only through the factor of age.
  • Misinterpretation of smallpox vaccine efficacy due to lack of deeper questioning.

3. The Ladder of Causation: Association as the First Step

The first rung of the Ladder of Causation emphasizes observation and probability. Humans and animals naturally form connections by observing the world. Machines like self-driving cars, however, struggle at this level because they lack a deeper understanding of relationships.

An owl tracks prey purely through movement—it doesn’t ask why the prey is running. Similarly, self-driving cars can't deduce a pedestrian’s behavior to honking; they need every potential response programmed.

Statistics aligns with this rung, as it’s built on observing probabilities, like assessing the link between toothpaste purchases and floss. Yet, while this information is helpful, it does not establish causes—only likelihoods.

Examples

  • Owls reacting to prey movement without questioning motivations.
  • Self-driving cars’ inability to predict pedestrian reaction to honking.
  • Marketing studies analyzing the probability of purchasing products like floss with toothpaste.

4. Intervention: Active Change Unlocks the Second Rung

The second rung involves actively intervening in the world. When humans take purposeful actions, like asking "What if we do this?" they move past observation. Controlled experiments embody this step by testing specific interventions.

Controlled trials date back to ancient history. Daniel, a Biblical figure, proved a vegetarian diet’s benefits through such an experiment with King Nebuchadnezzar’s boys. Modern examples include Facebook A/B testing to optimize webpage layouts.

Machines lag at this level because they can’t currently test cause-effect scenarios independently. Unlike humans, machines don’t think in terms of deliberate intervention—they only observe.

Examples

  • Daniel’s ancient dietary experiment for the Babylonian king.
  • Facebook’s layout tests to analyze user behavior.
  • Taking a painkiller to combat a headache as an everyday intervention.

5. Counterfactual Thinking: Envisioning Alternate Outcomes

Humans excel at imagining how different decisions might produce different outcomes. This kind of thinking resides on the ladder’s third rung, where humans ask "what if" when assessing cause and effect.

An example of counterfactual reasoning includes climate studies that ask, “What if CO2 levels were at preindustrial rates?” Legal proceedings also explore counterfactuals during investigations to evaluate the necessity of someone’s actions for an event.

Artificial Intelligence struggles here because it cannot generalize the way humans can. Complex questions—like whether oxygen or a match "caused" a fire—are difficult for machines to interpret without human-like reasoning.

Examples

  • Climate models exploring alternate CO2 levels.
  • Trials determining "but-for" causation, like inferring responsibility in a shooting.
  • Machines struggling with deciding the requisite cause in a fire scenario.

6. Confounders Complicate Causal Research

Experiments must account for variables known as confounders—hidden factors that influence both the cause and the effect. Ignoring these leads to false correlations and faulty conclusions.

Take the debate over smoking and lung cancer: opponents proposed genetic predisposition as a confounder during the 1950s. Later, controlled experiments and randomization debunked such arguments.

Randomization helps, but it can be unethical for some studies, like assigning smoking habits. Researchers have to instead creatively control for confounders to isolate true relationships.

Examples

  • Genetics as a proposed but debunked confounder in the smoking–lung cancer link.
  • Randomization helping uncover placebo impacts in clinical trials.
  • Controlled grouping avoiding income biases in drug trials.

7. Mediators Explain “Why” Causation Works

Mediators illustrate how one cause translates into its effect, making them essential. A fire triggers an alarm by releasing smoke, which acts as the mediator between cause and effect.

Misidentifying mediators can lead to disaster. For instance, early sailors mistook citrus acid for the key to combating scurvy. But later, when lime juice lacked vitamin C, entire crews faced horrific outbreaks.

Understanding mediators through counterfactual thinking allows effective intervention. If they’d known about vitamins, Arctic and Antarctic exploration history might look very different today.

Examples

  • Smoke acting as the link between fire and an alarm.
  • British Navy misattributing scurvy prevention to citrus acidity instead of vitamin C.
  • Arctic explorers failing to prevent scurvy due to improperly identified mediators.

8. Expressing Causation Mathematically

Relationships between factors can be mapped through causal diagrams, revealing mediators and confounders. These diagrams logically lay out factors and their interactions, preparing information for deeper analysis.

Creating formulas from diagrams allows predictions. For instance, researchers testing a blood-pressure drug could assess its impact relative to age as a confounder, quantifying the drug’s true effect.

Computers could benefit immensely from this method, allowing machines to finally ask not just “what happens?” but “why does this happen?”

Examples

  • Causal diagrams clarifying mediators in blood pressure tests.
  • Mathematical formulae measuring drug effects while accounting for age.
  • Computers breaking through observational barriers using programmed causal models.

9. The Future of AI and Causation

Teaching computers to process causation would revolutionize research. With algorithms able to generate “why” questions, machines could answer challenges that today require human expertise.

Imagine an AI that calculates the factors allowing life on distant planets. Or models figuring out which genes heighten cancer risks. With causality integrated, machines could address both abstract and practical challenges.

This new frontier sets up not only technological advancements but also quicker, richer scientific breakthroughs across disciplines.

Examples

  • AI assessing habitability criteria on exoplanets.
  • Computer-assisted cancer-genetics research targeting risk genes.
  • Machines generating counterfactual scenarios for climate-change interventions.

Takeaways

  1. Question surface-level data; look deeper at causes and relationships before drawing conclusions.
  2. Use causal diagrams and path models to clearly define and analyze relationships in studies or decision-making.
  3. Support advancements in AI research to integrate causal reasoning, revolutionizing problem-solving across industries.

Books like The Book of Why