Show simple item record

dc.contributor.authorChetverikov, Denis
dc.contributor.authorHansen, Christian
dc.contributor.authorRobins, James
dc.contributor.authorChernozhukov, Victor V
dc.contributor.authorDemirer, Mert
dc.contributor.authorDuflo, Esther
dc.contributor.authorNewey, Whitney K
dc.date.accessioned2018-03-01T21:22:20Z
dc.date.available2018-03-01T21:22:20Z
dc.date.issued2018-01
dc.date.submitted2016-10
dc.identifier.issn13684221
dc.identifier.urihttp://hdl.handle.net/1721.1/113906
dc.description.abstractWe revisit the classic semi-parametric problem of inference on a low-dimensional parameter θ 0 in the presence of high-dimensional nuisance parameters η 0 . We depart from the classical setting by allowing for η 0 to be so high-dimensional that the traditional assumptions (e.g. Donsker properties) that limit complexity of the parameter space for this object break down. To estimate η 0 , we consider the use of statistical or machine learning (ML) methods, which are particularly well suited to estimation in modern, very high-dimensional cases. ML methods perform well by employing regularization to reduce variance and trading off regularization bias with overfitting in practice. However, both regularization bias and overfitting in estimating η 0 cause a heavy bias in estimators of θ 0 that are obtained by naively plugging ML estimators of η 0 into estimating equations for θ 0 . This bias results in the naive estimator failing to be N-1/2 consistent, where N is the sample size. We show that the impact of regularization bias and overfitting on estimation of the parameter of interest θ 0 can be removed by using two simple, yet critical, ingredients: (1) using Neyman-orthogonal moments/scores that have reduced sensitivity with respect to nuisance parameters to estimate θ 0 ; (2) making use of cross-fitting, which provides an efficient form of data-splitting. We call the resulting set of methods double or debiased ML (DML). We verify that DML delivers point estimators that concentrate in an N-1/2-neighbourhood of the true parameter values and are approximately unbiased and normally distributed, which allows construction of valid confidence statements. The generic statistical theory of DML is elementary and simultaneously relies on only weak theoretical requirements, which will admit the use of a broad array of modern ML methods for estimating the nuisance parameters, such as random forests, lasso, ridge, deep neural nets, boosted trees, and various hybrids and ensembles of these methods. We illustrate the general theory by applying it to provide theoretical properties of the following: DML applied to learn the main regression parameter in a partially linear regression model; DML applied to learn the coefficient on an endogenous variable in a partially linear instrumental variables model; DML applied to learn the average treatment effect and the average treatment effect on the treated under unconfoundedness; DML applied to learn the local average treatment effect in an instrumental variables setting. In addition to these theoretical applications, we also illustrate the use of DML in three empirical examples.en_US
dc.description.sponsorshipNational Science Foundationen_US
dc.publisherWiley-Blackwellen_US
dc.relation.isversionofhttp://dx.doi.org/10.1111/ECTJ.12097en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleDouble/debiased machine learning for treatment and structural parametersen_US
dc.typeArticleen_US
dc.identifier.citationChernozhukov, Victor, Denis Chetverikov, Mert Demirer, Esther Duflo, Christian Hansen, Whitney Newey, and James Robins. “Double/debiased Machine Learning for Treatment and Structural Parameters.” The Econometrics Journal 21, no. 1 (January 16, 2018): C1–C68.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Economicsen_US
dc.contributor.departmentMassachusetts Institute of Technology. School of Humanities, Arts, and Social Sciencesen_US
dc.contributor.mitauthorChernozhukov, Victor V
dc.contributor.mitauthorDemirer, Mert
dc.contributor.mitauthorDuflo, Esther
dc.contributor.mitauthorNewey, Whitney K
dc.relation.journalThe Econometrics Journalen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2018-02-20T17:42:32Z
dspace.orderedauthorsChernozhukov, Victor; Chetverikov, Denis; Demirer, Mert; Duflo, Esther; Hansen, Christian; Newey, Whitney; Robins, Jamesen_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-3250-6714
dc.identifier.orcidhttps://orcid.org/0000-0003-4300-4258
dc.identifier.orcidhttps://orcid.org/0000-0001-6105-617X
dc.identifier.orcidhttps://orcid.org/0000-0003-2699-4704
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record