Causal ML in Healthcare: Beyond Prediction to Prevention

Every day, surgical departments face critical decisions that impact patient lives. Imagine a scenario: A hospital’s AI system predicts higher complications for morning surgeries, leading administrators to recommend moving all complex procedures to afternoon slots. But this decision, based on traditional machine learning, might actually harm patients. This article will explain why — and how a new approach called Causal Machine Learning (Causal ML) could transform healthcare decision-making for the better.

Risk prediction

There is a lot of excitement about AI-driven risk prediction. It has the potential to:

While traditional Machine Learning excels at finding patterns in medical data, it identifies correlations but does not determine true cause and effect. This limitation can lead to false and potentially dangerous conclusions in healthcare settings.

Consider our fictional hospital MorningStar. Their ML model accurately showed worse outcomes for morning surgeries. However, this wasn’t because morning surgeries were inherently less safe. Rather, the hospital policy was to schedule complex, high-risk cases in the morning when teams were fresh. The AI had discovered a correlation but missed the underlying causal relationship — leading to potentially misleading recommendations.

In actual fact, complexity affects both time-of-day, the treatment and the complication rate, the outcome. It is a confounder.

Causal graph showing the relationship between complexity, time-of-day, and outcomes

The model learnt to maximise accuracy by using time-of-day, and that works really well for MorningStar. But now imagine another hospital AllStar, where they schedule complex throughout the day to spread the load. The model trained at MorningStar will not generalise well to a new setting where the statistics are different. In ML terminology, we say that the model is ‘biassed’ and does not generalise ‘out-of-distribution’.

Limitations and opportunities

This example highlights three critical limitations of traditional (non-causal) ML in healthcare:

  1. It can’t distinguish between correlation and causation, leading to misleading conclusions
  2. It assumes historical patterns will hold true in new situations
  3. It can’t answer “what-if” questions about different treatment choices

Moreover, for AI to be trusted and useful, explainability is crucial — clinicians should understand ‘why’ a prediction was made. There are methods to ‘explain’ model predictions, but they can only explain the correlations learnt by the model and not the factors that are causally related to those outcomes.

Also, we can achieve more, by breaking out of a traditional mindset and going beyond straightforward risk prediction. The highest risk patients are not necessarily the ones whose outcomes we can improve. For example, they may be the sickest patients, without any modifiable factors. In order to intervene in the most effective way, it is better to ask which risks are most preventable, and which input-variables are modifiable and causally related to the outcome. This will enable healthcare systems to ensure sustainability by better allocating resources to have a higher impact.

Causal ML

Causal ML represents a fundamental shift in how we analyse healthcare data. Instead of simply identifying patterns, it helps us to understand the actual mechanisms behind patient outcomes, which unlocks opportunities that we’ll discuss. Think of it as upgrading from a traffic forecast that predicts congestion to one that explains why it’s going to be congested and what we can do to prepare or mitigate.

Causal ML consists of two main domains: Discovery and Inference.

Causal Discovery

Causal Inference

Causal inference can usually be done at two levels. On the group level (i.e. considering all the examples in the dataset), which could inform protocols and process, or individualised for personalised precision medicine.

More on ‘what-if’ scenarios

What if we could see how different treatment decisions will affect reality? Asking ‘what-if’ questions allows us to explore alternative scenarios using the data we already have (i.e. to ask counterfactual questions with only observational data). Traditionally, to understand whether a treatment really works, we’d need to run a randomised controlled trial (RCT) — the gold standard of medical research. But RCT’s are expensive and often not practical or ethical. For example, we can’t ask people to smoke, to study its effects on surgery. And if we are studying rare conditions like unusual tumours, gathering enough patients for a trial could take decades.

Real-world impact

Let’s revisit our surgery timing example through a Causal ML lens:

Distinguishing between correlation and causation

What-if scenarios / counterfactuals

More generalisable models

‘Preventable’ risk

The future of healthcare decision-making

While traditional ML has improved healthcare prediction, Causal ML could take us to the next level — from merely predicting outcomes to understanding and influencing them more effectively. This isn’t just about better predictions; it’s about empowering healthcare providers to make more informed, effective decisions that truly improve patient care.

Footnotes

  1. Discovering the correct causal graph is difficult as there are unknowns and hidden assumptions. So it’s good practice to derive multiple causal graphs and regard them as hypotheses which are challenged and refined.

  2. There are other important variable types such as mediators (the ‘freshness’ in this example) and colliders. Colliders are affected by both treatment and variable, and should not be controlled.

Want to see how Atidia can help your hospital?

Book a free, no-obligation pathway review with our clinical team.

Book a Review