Show simple item record

dc.contributor.authorHuggins, Jonathan H.
dc.contributor.authorCampbell, Trevor David
dc.contributor.authorBroderick, Tamara A
dc.date.accessioned2021-01-27T18:50:42Z
dc.date.available2021-01-27T18:50:42Z
dc.date.issued2016-12
dc.identifier.issn1049-5258
dc.identifier.urihttps://hdl.handle.net/1721.1/129582
dc.description.abstractThe use of Bayesian methods in large-scale data settings is attractive because of the rich hierarchical models, uncertainty quantification, and prior specification they provide. Standard Bayesian inference algorithms are computationally expensive, however, making their direct application to large datasets difficult or infeasible. Recent work on scaling Bayesian inference has focused on modifying the underlying algorithms to, for example, use only a random data subsample at each iteration. We leverage the insight that data is often redundant to instead obtain a weighted subset of the data (called a coreset) that is much smaller than the original dataset. We can then use this small coreset in any number of existing posterior inference algorithms without modification. In this paper, we develop an efficient coreset construction algorithm for Bayesian logistic regression models. We provide theoretical guarantees on the size and approximation quality of the coreset - both for fixed, known datasets, and in expectation for a wide class of data generative models. Crucially, the proposed approach also permits efficient construction of the coreset in both streaming and parallel settings, with minimal additional effort. We demonstrate the efficacy of our approach on a number of synthetic and real-world datasets, and find that, in practice, the size of the coreset is independent of the original dataset size. Furthermore, constructing the coreset takes a negligible amount of time compared to that required to run MCMC on it.en_US
dc.description.sponsorshipUnited States. Office of Naval Research. Multidisciplinary University Research Initiative (Grant N000141110688)en_US
dc.language.isoen
dc.publisherCurranen_US
dc.relation.isversionofhttps://papers.nips.cc/paper/2016/hash/2b0f658cbffd284984fb11d90254081f-Abstract.htmlen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceNeural Information Processing Systems (NIPS)en_US
dc.titleCoresets for scalable Bayesian logistic regressionen_US
dc.typeArticleen_US
dc.identifier.citationHuggins, Jonathan H. et al. “Coresets for scalable Bayesian logistic regression.” Paper presented at the 30th Conference on Neural Information Processing Systems (NIPS 2016), Bacelona, Spain, December 5-10 2016, Curran © 2016 The Author(s)en_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.relation.journal30th Conference on Neural Information Processing Systems (NIPS 2016)en_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2020-12-03T17:45:31Z
dspace.orderedauthorsHuggins, JH; Campbell, T; Broderick, Ten_US
dspace.date.submission2020-12-03T17:45:35Z
mit.journal.volume2016en_US
mit.licensePUBLISHER_POLICY
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record