Show simple item record

dc.contributor.authorAhmed, Asrar
dc.contributor.authorVarakantham, Pradeep
dc.contributor.authorAdulyasak, Yossiri
dc.contributor.authorJaillet, Patrick
dc.date.accessioned2015-12-19T02:23:48Z
dc.date.available2015-12-19T02:23:48Z
dc.date.issued2013
dc.identifier.issn1049-5258
dc.identifier.urihttp://hdl.handle.net/1721.1/100443
dc.description.abstractIn this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust optimization approaches for these problems have focussed on the computation of {\em maximin} policies which maximize the value corresponding to the worst realization of the uncertainty. Recent work has proposed {\em minimax} regret as a suitable alternative to the {\em maximin} objective for robust optimization. However, existing algorithms for handling {\em minimax} regret are restricted to models with uncertainty over rewards only. We provide algorithms that employ sampling to improve across multiple dimensions: (a) Handle uncertainties over both transition and reward models; (b) Dependence of model uncertainties across state, action pairs and decision epochs; (c) Scalability and quality bounds. Finally, to demonstrate the empirical effectiveness of our sampling approaches, we provide comparisons against benchmark algorithms on two domains from literature. We also provide a Sample Average Approximation (SAA) analysis to compute a posteriori error bounds.en_US
dc.description.sponsorshipSingapore. National Research Foundation (Singapore-MIT Alliance for Research and Technology Center. Future Urban Mobility Program)en_US
dc.description.sponsorshipUnited States. Office of Naval Research (Grant N00014-12-1-0999)en_US
dc.language.isoen_US
dc.publisherNeural Information Processing Systemsen_US
dc.relation.isversionofhttp://papers.nips.cc/paper/4970-regret-based-robust-solutions-for-uncertain-markov-decision-processesen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceNIPSen_US
dc.titleRegret Based Robust Solutions for Uncertain Markov Decision Processesen_US
dc.typeArticleen_US
dc.identifier.citationAhmed, Asrar, Pradeep Varakantham, Yossiri Adulyasak, and Patrick Jaillet. "Regret Based Robust Solutions for Uncertain Markov Decision Processes." Advances in Neural Information Processing Systems 26 (NIPS 2013).en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.mitauthorAdulyasak, Yossirien_US
dc.contributor.mitauthorJaillet, Patricken_US
dc.relation.journalAdvances in Neural Information Processing Systems (NIPS)en_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsAhmed, Asrar; Varakantham, Pradeep; Adulyasak, Yossiri; Jaillet, Patricken_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8585-6566
mit.licensePUBLISHER_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record