Arbeitspapier
Double machine learning for treatment and causal parameters
Most modern supervised statistical/machine learning (ML) methods are explicitly designed to solve prediction problems very well. Achieving this goal does not imply that these methods automatically deliver good estimators of causal parameters. Examples of such parameters include individual regression coefficients, average treatment effects, average lifts, and demand or supply elasticities. In fact, estimators of such causal parameters obtained via naively plugging ML estimators into estimating equations for such parameters can behave very poorly. For example, the resulting estimators may formally have inferior rates of convergence with respect to the sample size n caused by regularization bias. Fortunately, this regularization bias can be removed by solving auxiliary prediction problems via ML tools. Specifically, we can form an efficient score for the target low-dimensional parameter by combining auxiliary and main ML predictions. The efficient score may then be used to build an efficient estimator of the target parameter which typically will converge at the fastest possible 1/Í n rate and be approximately unbiased and normal, allowing simple construction of valid confidence intervals for parameters of interest. The resulting method thus could be called a "double ML" method because it relies on estimating primary and auxiliary predictive models. Such double ML estimators achieve the fastest rates of convergence and exhibit robust good behavior with respect to a broader class of probability distributions than naive "single" ML estimators. In order to avoid overfitting, following [3], our construction also makes use of the K-fold sample splitting, which we call cross-fitting. The use of sample splitting allows us to use a very broad set of ML predictive methods in solving the auxiliary and main prediction problems, such as random forests, lasso, ridge, deep neural nets, boosted trees, as well as various hybrids and aggregates of these methods (e.g. a hybrid of a random forest and lasso). We illustrate the application of the general theory through application to the leading cases of estimation and inference on the main parameter in a partially linear regression model and estimation and inference on average treatment effects and average treatment effects on the treated under conditional random assignment of the treatment. These applications cover randomized control trials as a special case. We then use the methods in an empirical application which estimates the effect of 401(k) eligibility on accumulated financial assets.
- Language
-
Englisch
- Bibliographic citation
-
Series: cemmap working paper ; No. CWP49/16
- Classification
-
Wirtschaft
- Subject
-
Neyman
Orthogonalization
cross-fit
double machine learning
debiased machine learning
orthogonal score
efficient score
post-machine-learning and post-regularization inference
random forest
lasso
deep learning
neural nets
boosted trees
efficiency
optimality
- Event
-
Geistige Schöpfung
- (who)
-
Chernozhukov, Victor
Chetverikov, Denis
Demirer, Mert
Duflo, Esther
Hansen, Christian
Newey, Whitney K.
- Event
-
Veröffentlichung
- (who)
-
Centre for Microdata Methods and Practice (cemmap)
- (where)
-
London
- (when)
-
2016
- DOI
-
doi:10.1920/wp.cem.2016.4916
- Handle
- Last update
-
10.03.2025, 11:43 AM CET
Data provider
ZBW - Deutsche Zentralbibliothek für Wirtschaftswissenschaften - Leibniz-Informationszentrum Wirtschaft. If you have any questions about the object, please contact the data provider.
Object type
- Arbeitspapier
Associated
- Chernozhukov, Victor
- Chetverikov, Denis
- Demirer, Mert
- Duflo, Esther
- Hansen, Christian
- Newey, Whitney K.
- Centre for Microdata Methods and Practice (cemmap)
Time of origin
- 2016