Show simple item record

dc.contributor.authorJordan, MI
dc.contributor.authorStephenson, William T.(William Thomas)
dc.contributor.authorBroderick, Tamara A
dc.date.accessioned2020-12-10T16:36:37Z
dc.date.available2020-12-10T16:36:37Z
dc.date.issued2019-04
dc.identifier.issn1938-7228
dc.identifier.urihttps://hdl.handle.net/1721.1/128774
dc.description.abstractThe error or variability of machine learning algorithms is often assessed by repeatedly refitting a model with different weighted versions of the observed data. The ubiquitous tools of cross-validation (CV) and the bootstrap are examples of this technique. These methods are powerful in large part due to their model agnosticism but can be slow to run on modern, large data sets due to the need to repeatedly re-fit the model. In this work, we use a linear approximation to the dependence of the fitting procedure on the weights, producing results that can be faster than repeated re-fitting by an order of magnitude. This linear approximation is sometimes known as the “infinitesimal jackknife” in the statistics literature, where it is mostly used as a theoretical tool to prove asymptotic results. We provide explicit finite-sample error bounds for the infinitesimal jackknife in terms of a small number of simple, verifiable assumptions. Our results apply whether the weights and data are stochastic or deterministic, and so can be used as a tool for proving the accuracy of the infinitesimal jackknife on a wide variety of problems. As a corollary, we state mild regularity conditions under which our approximation consistently estimates true leave-k-out cross-validation for any fixed k. These theoretical results, together with modern automatic differentiation software, support the application of the infinitesimal jackknife to a wide variety of practical problems in machine learning, providing a “Swiss Army infinitesimal jackknife.” We demonstrate the accuracy of our methods on a range of simulated and real datasets.en_US
dc.description.sponsorshipUnited States. Defense Advanced Research Projects Agency (Grant A8650-18-2-7832)en_US
dc.language.isoen
dc.publisherPMLRen_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleA Swiss army infinitesimal jackknifeen_US
dc.typeArticleen_US
dc.identifier.citationGiordano, Ryan et al. “A Swiss army infinitesimal jackknife.” Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics (AISTATS), 89 (April 2019): © 2019 The Author(s)en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.relation.journalProceedings of the 22ndInternational Conference on Ar-tificial Intelligence and Statistics (AISTATS)en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2020-12-03T18:38:20Z
dspace.orderedauthorsGiordano, R; Stephenson, W; Liu, R; Jordan, MI; Broderick, Ten_US
dspace.date.submission2020-12-03T18:38:29Z
mit.journal.volume89en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record