Learning with Deictic Representation
dc.contributor.author | Finney, Sarah | en_US |
dc.contributor.author | Gardiol, Natalia H. | en_US |
dc.contributor.author | Kaelbling, Leslie Pack | en_US |
dc.contributor.author | Oates, Tim | en_US |
dc.date.accessioned | 2004-10-08T20:37:45Z | |
dc.date.available | 2004-10-08T20:37:45Z | |
dc.date.issued | 2002-04-10 | en_US |
dc.identifier.other | AIM-2002-006 | en_US |
dc.identifier.uri | http://hdl.handle.net/1721.1/6685 | |
dc.description.abstract | Most reinforcement learning methods operate on propositional representations of the world state. Such representations are often intractably large and generalize poorly. Using a deictic representation is believed to be a viable alternative: they promise generalization while allowing the use of existing reinforcement-learning methods. Yet, there are few experiments on learning with deictic representations reported in the literature. In this paper we explore the effectiveness of two forms of deictic representation and a naive propositional representation in a simple blocks-world domain. We find, empirically, that the deictic representations actually worsen performance. We conclude with a discussion of possible causes of these results and strategies for more effective learning in domains with objects. | en_US |
dc.format.extent | 41 p. | en_US |
dc.format.extent | 5712208 bytes | |
dc.format.extent | 1294450 bytes | |
dc.format.mimetype | application/postscript | |
dc.format.mimetype | application/pdf | |
dc.language.iso | en_US | |
dc.relation.ispartofseries | AIM-2002-006 | en_US |
dc.subject | AI | en_US |
dc.subject | Reinforcement Learning | en_US |
dc.subject | Partial Observability | en_US |
dc.subject | Representations | en_US |
dc.title | Learning with Deictic Representation | en_US |