Skip to content

Latest commit

 

History

History
36 lines (22 loc) · 1.3 KB

README.md

File metadata and controls

36 lines (22 loc) · 1.3 KB

Follow the .ipynb file for the demo

Original Paper by:

Double/Debiased Machine Learning for Treatment and Causal Parameters

Victor Chernozhukov, Denis Chetverikov, Mert Demirer, Esther Duflo, Christian Hansen, Whitney Newey, James Robins (2017)

Link to Paper
Link to documentation

Additional Resources:

  1. DoubleML in Python

Example from "Be careful when interpreting predictive models in search of causal insights."


A joint article about causality and interpretable machine learning with Eleanor Dillon, Jacob LaRiviere, Scott Lundberg, Jonathan Roth, and Vasilis Syrgkanis from Microsoft.


Predictive models e.g. XGBoost coupled with ML Interpretability models e.g. SHAP are powerful. But they are only useful for:

  1. Predictions
  2. Relationship between inputs and outcomes
  3. Diagnosis of potential problems

Predictive models should not be used for 'decision making'. Since predictive models are not causal!! Predictive models implicitly assume that everyone will keep behaving the same way in the future, and therefore correlation patterns will stay constant. But they do not model behavior.