Show simple item record

dc.contributor.authorAdulyasak, Yossiri
dc.contributor.authorVarakantham, Pradeep
dc.contributor.authorAhmed, Asrar
dc.contributor.authorJaillet, Patrick
dc.date.accessioned2018-06-12T13:26:47Z
dc.date.available2018-06-12T13:26:47Z
dc.date.issued2015-01
dc.identifier.isbnISBN:0-262-51129-0
dc.identifier.urihttp://hdl.handle.net/1721.1/116234
dc.description.abstractMarkov Decision Problems, MDPs offer an effective mechanism for planning under uncertainty. However, due to unavoidable uncertainty over models, it is difficult to obtain an exact specification of an MDP. We are interested in solving MDPs, where transition and reward functions are not exactly specified. Existing research has primarily focussed on computing infinite horizon stationary policies when optimizing robustness, regret and percentile based objectives. We focus specifically on finite horizon problems with a special emphasis on objectives that are separable over individual instantiations of model uncertainty (i.e., objectives that can be expressed as a sum over instantiations of model uncertainty): (a) First, we identify two separable objectives for uncertain MDPs: Average Value Maximization (AVM) and Confidence Probability Maximisation (CPM). (b) Second, we provide optimization based solutions to compute policies for uncertain MDPs with such objectives. In particular, we exploit the separability of AVM and CPM objectives by employing Lagrangian dual decomposition (LDD). (c) Finally, we demonstrate the utility of the LDD approach on a benchmark problem from the literature.en_US
dc.description.sponsorshipNational Research Foundation of Singaporeen_US
dc.language.isoen_US
dc.publisherAAAI Pressen_US
dc.relation.isversionofhttp://dl.acm.org/citation.cfm?id=2888196en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceMIT Web Domainen_US
dc.titleSolving uncertain MDPs with objectives that are separable over instantiations of model uncertaintyen_US
dc.typeArticleen_US
dc.identifier.citationAdulyasak, Yossiri, Pradeep Varakantham, Asrar Ahmed and Patrick Jaillet. "Solving uncertain MDPs with objectives that are separable over instantiations of model uncertainty." In Proceeding AAAI'15 Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, Texas, January 25-30 2015, AAAI Press, ©2015, pp. 3454-3460.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Civil and Environmental Engineeringen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.mitauthorAdulyasak, Yossiri
dc.contributor.mitauthorJaillet, Patrick
dc.relation.journalProceeding AAAI'15 Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligenceen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsAdulyasak, Yossiri; Varakantham, Pradeep; Ahmed, Asrar; Jaillet, Patricken_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8585-6566
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record