An Interpretable Stroke Prediction Model using Rules and Bayesian Analysis
Author(s)
Letham, Benjamin; Rudin, Cynthia; McCormick, Tyler H.; Madigan, David
DownloadRudin_SWP_5040-13.pdf (474.1Kb)
Terms of use
Metadata
Show full item recordAbstract
We aim to produce predictive models that are not only accurate, but are also interpretable to human experts. Our models are decision lists, which consist of a series of if...then... statements (for example, if high blood pressure, then stroke) that discretize a high-dimensional, multivariate feature space into a series of simple, readily inter-
pretable decision statements. We introduce a generative model called the Bayesian List Machine which yields a posterior distribution over possible decision lists. It employs a novel prior structure to encourage sparsity. Our experiments show that the Bayesian List Machine has predictive accuracy on par with the current top algorithms
for prediction in machine learning. Our method is motivated by recent developments in personalized medicine, and can be used to produce highly accurate and interpretable medical scoring systems. We demonstrate this by producing an alternative to the CHADS2 score, actively used in clinical practice for estimating the risk of stroke in patients that have atrial brillation. Our model is as interpretable as CHADS2, but more accurate.
Date issued
2013-11-15Series/Report no.
MIT Sloan School of Management Working Paper;5040-13
Collections
The following license files are associated with this item: