Big Picture of Debiased Machine Learning
Debiased machine learning (DML) is a generic recipe. The idea behind it is adding a correction term to the plug-in estimator of the functional, which leads to properties such as semi-parametric efficiency, double robustness, and Neyman orthogonality.

(Auto)-DML is a Method-of-Moments estimator
-
debiased/orthognal moment scores
-
Why it matters?
- try to solve: model selection and/or regularization bias from ML learners (e.g. Lasso)
- Neyman orthogonality: ensure the parameter of interest insentitive to first order perturbation of nuisance estimation
- double robustness
- asymptotic normality
-
Key Idea: Debiasing is achieved by adding a correction term to the plug-in estimator of the functional
-
Three representations:
, where is outcome regression; is Rieze Representer (RR); is a contious linear functional; is data containing treatment and covariates ; is potential outcome
-
Correct the residual using RR
-
-
How to construct orthogonal moment function?
-
orthogonal moment function = identifying moment function + first step influence function (FSIF)
-
identifying moment function:
- involving outcome regression
-
FSIF:
- correct the residual using Rieze Representer (RR)
- Rieze Representer (RR)
- In the case of ATE with binary treatment, RR are inverse propensity score terms
- RR can be automatically characterized; NO NEED to know its analytical form
- Can use random forests and NNet learners of RR
-
-
Double Robustness
-
-
The score will be zero in expectation when either
or
-
-
-
Cross-fitting
- Why it matters?
- Reduce overfitting bias
- Why it matters?