Approximate dynamic programming using model-free Bellman Residual Elimination
Author(s)Bethke, Brett M.; How, Jonathan P.
DownloadHow-2010-Approximate dynamic programming using model-free Bellman Residual Elimination.pdf (296.7Kb)
MetadataShow full item record
This paper presents an modification to the method of Bellman Residual Elimination (BRE) for approximate dynamic programming. While prior work on BRE has focused on learning an approximate policy for an underlying Markov Decision Process (MDP) when the state transition model of the MDP is known, this work proposes a model-free variant of BRE that does not require knowledge of the state transition model. Instead, state trajectories of the system, generated using simulation and/or observations of the real system in operation, are used to build stochastic approximations of the quantities needed to carry out the BRE algorithm. The resulting algorithm can be shown to converge to the policy produced by the nominal, model-based BRE algorithm in the limit of observing an infinite number of trajectories. To validate the performance of the approach, we compare model-based and model-free BRE against LSPI, a well-known approximate dynamic programming algorithm. Measuring performance in terms of both computational complexity and policy quality, we present results showing that BRE performs at least as well as, and sometimes significantly better than, LSPI on a standard benchmark problem.
DepartmentMassachusetts Institute of Technology. Aerospace Controls Laboratory; Massachusetts Institute of Technology. Department of Aeronautics and Astronautics
American Control Conference 2010
Institute of Electrical and Electronics Engineers / American Automatic Control Council
Bethke, B., and J.P. How. “Approximate dynamic programming using model-free Bellman Residual Elimination.” American Control Conference (ACC), 2010. 2010. 4146-4151. Print.
Final published version
INSPEC Accession Number: 11508712