Show simple item record

dc.contributor.authorHazimeh, Hussein
dc.contributor.authorMazumder, Rahul
dc.date.accessioned2021-04-08T15:34:56Z
dc.date.available2021-04-08T15:34:56Z
dc.date.issued2020-08
dc.identifier.issn0030-364X
dc.identifier.issn1526-5463
dc.identifier.urihttps://hdl.handle.net/1721.1/130416
dc.description.abstractThe L₀-regularized least squares problem (a.k.a. best subsets) is central to sparse statistical learning and has attracted significant attention across the wider statistics, machine learning, and optimization communities. Recent work has shown that modern mixed integer optimization (MIO) solvers can be used to address small to moderate instances of this problem. In spite of the usefulness of L₀-based estimators and generic MIO solvers, there is a steep computational price to pay when compared with popular sparse learning algorithms (e.g., based on L₀ regularization). In this paper, we aim to push the frontiers of computation for a family of L₀-regularized problems with additional convex penalties. We propose a new hierarchy of necessary optimality conditions for these problems. We develop fast algorithms, based on coordinate descent and local combinatorial optimization, that are guaranteed to converge to solutions satisfying these optimality conditions. From a statistical viewpoint, an interesting story emerges. When the signal strength is high, our combinatorial optimization algorithms have an edge in challenging statistical settings. When the signal is lower, pure L₀ benefits from additional convex regularization. We empirically demonstrate that our family of L₀-based estimators can outperform the state-of-the-art sparse learning algorithms in terms of a combination of prediction, estimation, and variable selection metrics under various regimes (e.g., different signal strengths, feature correlations, number of samples and features). Our new open-source sparse learning toolkit L0Learn (available on CRAN and GitHub) reaches up to a threefold speedup (with p up to 10 ) when compared with competing toolkits such as glmnet and ncvreg. 0 0 1 0 0 0 6en_US
dc.description.sponsorshipUnited States. Office of Naval Research (ONR-N000141512342, ONR-N000141812298 (Young Investigator Award))en_US
dc.description.sponsorshipNational Science Foundation (U.S.) (Grant NSF-IIS-1718258)en_US
dc.language.isoen
dc.publisherInstitute for Operations Research and the Management Sciences (INFORMS)en_US
dc.relation.isversionof10.1287/OPRE.2019.1919en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleFast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithmsen_US
dc.typeArticleen_US
dc.identifier.citationHazimeh, Hussein and Rahul Mazumder. “Fast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithms.” Operations Research, 68, 5 (August 2020): iii-vi, 1285-1624, C2-C3 © 2020 The Author(s)en_US
dc.contributor.departmentMassachusetts Institute of Technology. Operations Research Centeren_US
dc.contributor.departmentSloan School of Managementen_US
dc.relation.journalOperations Researchen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2021-04-08T14:28:24Z
dspace.orderedauthorsHazimeh, H; Mazumder, Ren_US
dspace.date.submission2021-04-08T14:28:25Z
mit.journal.volume68en_US
mit.journal.issue5en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record