Causal ML
Methods that combine flexible machine learning with causal identification — enabling valid effect estimation in high-dimensional settings and principled estimation of treatment effect heterogeneity.
Classical causal methods require the researcher to specify control variables parametrically. In high-dimensional settings this introduces omitted variable bias and model misspecification.
ML models can approximate complex nuisance functions — the conditional mean of Y or T given X — without specifying a parametric form. But naive ML estimates are regularization-biased for causal parameters.
Causal ML resolves this tension: use flexible ML for the parts you don't care about (nuisance), and apply orthogonalization and cross-fitting to recover valid, root-n consistent causal estimates.
library(DoubleML)
# Partially linear regression via DML
obj <- DoubleMLPLR$new(
data = dml_data,
ml_l = lrn("regr.ranger"), # outcome model
ml_m = lrn("regr.ranger"), # treatment model
n_folds = 5,
score = "partialling out"
)
obj$fit()
obj$summary()Double machine learning
Uses cross-fitted ML models to partial out controls from both treatment and outcome, then estimates the causal effect on the residuals. Removes regularization bias without sacrificing flexibility.
Causal forest
Extends random forests to estimate heterogeneous treatment effects (CATE) at the individual level. Uses honest splitting and local centering to debias estimates across the covariate space.
AIPW / DR learner
Augmented inverse probability weighting combines outcome and propensity models into a doubly robust score — consistent if either nuisance model is correctly specified. Efficient semiparametrically.
Heterogeneous treatment effects
A family of methods — X-learner, R-learner, BART — for estimating how treatment effects vary across subgroups or individual covariates. Requires careful validation to avoid overfitting.
Targeted maximum likelihood
A doubly-robust, one-step efficient estimator that updates an initial ML outcome model via a targeting step solving the efficient score equation. Respects model bounds and achieves semiparametric efficiency when nuisance models converge.