dc.contributor.author | Shah, Julie A | |
dc.contributor.author | Nikolaidis, Stefanos | |
dc.contributor.author | Ramakrishnan, Ramya | |
dc.contributor.author | Gu, Keren | |
dc.date.accessioned | 2017-04-05T20:03:20Z | |
dc.date.available | 2017-04-05T20:03:20Z | |
dc.date.issued | 2015-03 | |
dc.identifier.isbn | 9781450328838 | |
dc.identifier.uri | http://hdl.handle.net/1721.1/107887 | |
dc.description.abstract | We present a framework for automatically learning human user models from joint-action demonstrations that enables a robot to compute a robust policy for a collaborative task with a human. First, the demonstrated action sequences are clustered into different human types using an unsupervised learning algorithm. A reward function is then learned for each type through the employment of an inverse reinforcement learning algorithm. The learned model is then incorporated into a mixed-observability Markov decision process (MOMDP) formulation, wherein the human type is a partially observable variable. With this framework, we can infer online the human type of a new user that was not included in the training set, and can compute a policy for the robot that will be aligned to the preference of this user. In a human subject experiment (n=30), participants agreed more strongly that the robot anticipated their actions when working with a robot incorporating the proposed framework (p<0.01), compared to manually annotating robot actions. In trials where participants faced difficulty annotating the robot actions to complete the task, the proposed framework significantly improved team efficiency (p<0.01). The robot incorporating the framework was also found to be more responsive to human actions compared to policies computed using a hand-coded reward function by a domain expert (p<0.01). These results indicate that learning human user models from joint-action demonstrations and encoding them in a MOMDP formalism can support effective teaming in human-robot collaborative tasks. | en_US |
dc.language.iso | en_US | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | en_US |
dc.relation.isversionof | http://dx.doi.org/10.1145/2696454.2696455 | en_US |
dc.rights | Creative Commons Attribution-Noncommercial-Share Alike | en_US |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/4.0/ | en_US |
dc.source | MIT web domain | en_US |
dc.title | Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Nikolaidis, Stefanos et al. “Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks.” ACM Press, 2015. 189–196. | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
dc.contributor.mitauthor | Shah, Julie A | |
dc.contributor.mitauthor | Nikolaidis, Stefanos | |
dc.contributor.mitauthor | Ramakrishnan, Ramya | |
dc.contributor.mitauthor | Gu, Keren | |
dc.relation.journal | Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction - HRI '15 | en_US |
dc.eprint.version | Author's final manuscript | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dspace.orderedauthors | Nikolaidis, Stefanos; Ramakrishnan, Ramya; Gu, Keren; Shah, Julie | en_US |
dspace.embargo.terms | N | en_US |
dc.identifier.orcid | https://orcid.org/0000-0003-1338-8107 | |
dc.identifier.orcid | https://orcid.org/0000-0001-8239-5963 | |
mit.license | OPEN_ACCESS_POLICY | en_US |