Show simple item record

dc.contributor.authorBandi, Hari
dc.contributor.authorBertsimas, Dimitris
dc.contributor.authorMazumder, Rahul
dc.date.accessioned2021-10-27T20:35:50Z
dc.date.available2021-10-27T20:35:50Z
dc.date.issued2019
dc.identifier.urihttps://hdl.handle.net/1721.1/136540
dc.description.abstract<jats:p> We consider the problem of estimating the parameters of a multivariate Gaussian mixture model (GMM) given access to n samples that are believed to have come from a mixture of multiple subpopulations. State-of-the-art algorithms used to recover these parameters use heuristics to either maximize the log-likelihood of the sample or try to fit first few moments of the GMM to the sample moments. In contrast, we present here a novel mixed-integer optimization (MIO) formulation that optimally recovers the parameters of the GMM by minimizing a discrepancy measure (either the Kolmogorov–Smirnov or the total variation distance) between the empirical distribution function and the distribution function of the GMM whenever the mixture component weights are known. We also present an algorithm for multidimensional data that optimally recovers corresponding means and covariance matrices. We show that the MIO approaches are practically solvable for data sets with n in the tens of thousands in minutes and achieve an average improvement of 60%–70% and 50%–60% on mean absolute percentage error in estimating the means and the covariance matrices, respectively, over the expectation–maximization (EM) algorithm independent of the sample size n. As the separation of the Gaussians decreases and, correspondingly, the problem becomes more difficult, the edge in performance in favor of the MIO methods widens. Finally, we also show that the MIO methods outperform the EM algorithm with an average improvement of 4%–5% on the out-of-sample accuracy for real-world data sets. </jats:p>
dc.language.isoen
dc.publisherInstitute for Operations Research and the Management Sciences (INFORMS)
dc.relation.isversionof10.1287/IJOO.2018.0009
dc.rightsCreative Commons Attribution-Noncommercial-Share Alike
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/
dc.sourceOther repository
dc.titleLearning a Mixture of Gaussians via Mixed-Integer Optimization
dc.typeArticle
dc.contributor.departmentMassachusetts Institute of Technology. Operations Research Center
dc.contributor.departmentSloan School of Management
dc.relation.journalINFORMS Journal on Optimization
dc.eprint.versionOriginal manuscript
dc.type.urihttp://purl.org/eprint/type/JournalArticle
eprint.statushttp://purl.org/eprint/status/NonPeerReviewed
dc.date.updated2021-02-05T19:33:00Z
dspace.orderedauthorsBandi, H; Bertsimas, D; Mazumder, R
dspace.date.submission2021-02-05T19:33:02Z
mit.journal.volume1
mit.journal.issue3
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Needed


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record