Show simple item record

dc.contributor.authorBertsimas, Dimitris
dc.contributor.authorDunn, Jack William
dc.date.accessioned2017-06-27T18:54:24Z
dc.date.available2018-02-04T06:00:05Z
dc.date.issued2017-04
dc.identifier.issn0885-6125
dc.identifier.issn1573-0565
dc.identifier.urihttp://hdl.handle.net/1721.1/110328
dc.description.abstractState-of-the-art decision tree methods apply heuristics recursively to create each split in isolation, which may not capture well the underlying characteristics of the dataset. The optimal decision tree problem attempts to resolve this by creating the entire decision tree at once to achieve global optimality. In the last 25 years, algorithmic advances in integer optimization coupled with hardware improvements have resulted in an astonishing 800 billion factor speedup in mixed-integer optimization (MIO). Motivated by this speedup, we present optimal classification trees, a novel formulation of the decision tree problem using modern MIO techniques that yields the optimal decision tree for axes-aligned splits. We also show the richness of this MIO formulation by adapting it to give optimal classification trees with hyperplanes that generates optimal decision trees with multivariate splits. Synthetic tests demonstrate that these methods recover the true decision tree more closely than heuristics, refuting the notion that optimal methods overfit the training data. We comprehensively benchmark these methods on a sample of 53 datasets from the UCI machine learning repository. We establish that these MIO methods are practically solvable on real-world datasets with sizes in the 1000s, and give average absolute improvements in out-of-sample accuracy over CART of 1–2 and 3–5% for the univariate and multivariate cases, respectively. Furthermore, we identify that optimal classification trees are likely to outperform CART by 1.2–1.3% in situations where the CART accuracy is high and we have sufficient training data, while the multivariate version outperforms CART by 4–7% when the CART accuracy or dimension of the dataset is low.en_US
dc.publisherSpringer USen_US
dc.relation.isversionofhttp://dx.doi.org/10.1007/s10994-017-5633-9en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceSpringer USen_US
dc.titleOptimal classification treesen_US
dc.typeArticleen_US
dc.identifier.citationBertsimas, Dimitris, and Jack Dunn. “Optimal Classification Trees.” Machine Learning 106.7 (2017): 1039–1082.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Operations Research Centeren_US
dc.contributor.mitauthorDunn, Jack William
dc.relation.journalMachine Learningen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2017-06-23T03:51:27Z
dc.language.rfc3066en
dc.rights.holderThe Author(s)
dspace.orderedauthorsBertsimas, Dimitris; Dunn, Jacken_US
dspace.embargo.termsNen
dc.identifier.orcidhttps://orcid.org/0000-0002-6936-4502
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record