Sloan Working Papers
http://hdl.handle.net/1721.1/1792
Wed, 23 Mar 2016 12:16:28 GMT2016-03-23T12:16:28ZSloan Working Papershttp://dspace.mit.edu:80/bitstream/id/5829/Sloan_201.pcsm.gif
http://hdl.handle.net/1721.1/1792
Technology Readiness Levels at 40: a study of state-of-the-art use, challenges, and opportunities
http://hdl.handle.net/1721.1/96307
Technology Readiness Levels at 40: a study of state-of-the-art use, challenges, and opportunities
Olechowski, Alison; Eppinger, Steven D.; Joglekar, Nitin
The technology readiness level (TRL) scale was introduced by NASA in the 1970s as a tool for assessing the maturity of technologies during complex system development. TRL data have been used to make multi-million dollar technology management decisions in programs such as NASA's Mars Curiosity Rover. This scale is now a de facto standard used for technology assessment and oversight in many industries, from power systems to consumer electronics. Low TRLs have been associated with significantly reduced timeliness and increased costs across a portfolio of US Department of Defense programs. However, anecdotal evidence raises concerns about many of the practices related to TRLs. We study TRL implementations based on semi-structured interviews with employees from seven different organizations and examine documentation collected from industry standards and organizational guidelines related to technology development and demonstration. Our findings consist of 15 challenges observed in TRL implementations that fall into three different categories: system complexity, planning and review, and validity of assessment. We explore research opportunities for these challenges and posit that addressing these opportunities, either singly or in groups, could improve decision processes and performance outcomes in complex engineering projects.
Wed, 01 Apr 2015 00:00:00 GMThttp://hdl.handle.net/1721.1/963072015-04-01T00:00:00ZThe Big Data Newsvendor: Practical Insights from Machine Learning
http://hdl.handle.net/1721.1/85658
The Big Data Newsvendor: Practical Insights from Machine Learning
Rudin, Cynthia; Vahn, Gah-Yi
We investigate the newsvendor problem when one has n observations of p features related to the demand as well as past demands. Both small data (p=n = o(1)) and big data (p=n = O(1)) are considered. For both cases, we propose a machine learning algorithm to solve the problem and derive a tight generalization bound on the expected out-of-sample cost. The algorithms can be extended intuitively to other situations, such as having censored demand data, ordering for multiple, similar items and having a new item with
limited data. We show analytically that our custom-designed, feature-based approach can be better than other data-driven approaches such as Sample Average Approximation (SAA) and separated estimation and optimization (SEO). Our method can also naturally incorporate the operational statistics method. We then apply the algorithms to nurse staffing in a hospital emergency room and show that (i) they can reduce the median out-of-sample cost by up to 46% and 16% compared to SAA and SEO respectively, with statistical significance at 0.01, and (ii) this is achieved either by carefully selecting a small number of features and applying the small data algorithm, or by using a large number of features and using the big data algorithm,
which automates feature-selection.
This is a revision of previously published DSpace entry: http://hdl.handle.net/1721.1/81412.
Thu, 06 Feb 2014 00:00:00 GMThttp://hdl.handle.net/1721.1/856582014-02-06T00:00:00ZAn Interpretable Stroke Prediction Model using Rules and Bayesian Analysis
http://hdl.handle.net/1721.1/82148
An Interpretable Stroke Prediction Model using Rules and Bayesian Analysis
Letham, Benjamin; Rudin, Cynthia; McCormick, Tyler H.; Madigan, David
We aim to produce predictive models that are not only accurate, but are also interpretable to human experts. Our models are decision lists, which consist of a series of if...then... statements (for example, if high blood pressure, then stroke) that discretize a high-dimensional, multivariate feature space into a series of simple, readily inter-
pretable decision statements. We introduce a generative model called the Bayesian List Machine which yields a posterior distribution over possible decision lists. It employs a novel prior structure to encourage sparsity. Our experiments show that the Bayesian List Machine has predictive accuracy on par with the current top algorithms
for prediction in machine learning. Our method is motivated by recent developments in personalized medicine, and can be used to produce highly accurate and interpretable medical scoring systems. We demonstrate this by producing an alternative to the CHADS2 score, actively used in clinical practice for estimating the risk of stroke in patients that have atrial brillation. Our model is as interpretable as CHADS2, but more accurate.
Fri, 15 Nov 2013 00:00:00 GMThttp://hdl.handle.net/1721.1/821482013-11-15T00:00:00ZThe Big Data Newsvendor: Practical Insights from Machine Learning Analysis
http://hdl.handle.net/1721.1/81412
The Big Data Newsvendor: Practical Insights from Machine Learning Analysis
Rudin, Cynthia; Vahn, Gah-Yi
We present a version of the newsvendor problem where one has n observations of p features as well as past demand. We consider both \big data" (p=n = O(1)) as well as small data (p=n = o(1)). For small data, we provide a linear programming machine learning algorithm that yields an asymptotically optimal order quantity. We also derive a generalization bound based on algorithmic stability, which is an upper bound on the expected out-of-sample cost. For big data, we propose a regularized version of the algorithm to address the curse of dimensionality. A generalization bound is derived for this case as well, bounding the out-of-sample cost with a quantity that depends on n and the amount of regularization. We apply the algorithm to analyze the newsvendor cost of nurse sta_ng using data from the emergency room of a large teaching hospital and show that (i) incorporating appropriate features can reduce the out-of-sample cost by up to 23% relative to the featureless Sample Average Approximation approach, and (ii) regularization can automate feature-selection while controlling the out-of-sample cost. By an appropriate choice of the newsvendor underage and overage costs, our results also apply to quantile regression.
A 2/6/2014 revision to this paper is available at http://hdl.handle.net/1721.1/85658.
Wed, 16 Oct 2013 00:00:00 GMThttp://hdl.handle.net/1721.1/814122013-10-16T00:00:00Z