dc.contributor.author | Chen, Yu Fan | |
dc.contributor.author | Liu, Miao | |
dc.contributor.author | Everett, Michael F | |
dc.contributor.author | How, Jonathan P | |
dc.date.accessioned | 2018-04-13T18:42:36Z | |
dc.date.available | 2018-04-13T18:42:36Z | |
dc.date.issued | 2017-07 | |
dc.identifier.isbn | 978-1-5090-4633-1 | |
dc.identifier.issn | 978-1-5090-4634-8 | |
dc.identifier.uri | http://hdl.handle.net/1721.1/114720 | |
dc.description.abstract | Finding feasible, collision-free paths for multiagent systems can be challenging, particularly in non-communicating scenarios where each agent's intent (e.g. goal) is unobservable to the others. In particular, finding time efficient paths often requires anticipating interaction with neighboring agents, the process of which can be computationally prohibitive. This work presents a decentralized multiagent collision avoidance algorithm based on a novel application of deep reinforcement learning, which effectively offloads the online computation (for predicting interaction patterns) to an offline learning procedure. Specifically, the proposed approach develops a value network that encodes the estimated time to the goal given an agent's joint configuration (positions and velocities) with its neighbors. Use of the value network not only admits efficient (i.e., real-time implementable) queries for finding a collision-free velocity vector, but also considers the uncertainty in the other agents' motion. Simulation results show more than 26% improvement in paths quality (i.e., time to reach the goal) when compared with optimal reciprocal collision avoidance (ORCA), a state-of-the-art collision avoidance strategy. | en_US |
dc.description.sponsorship | Ford Motor Company | en_US |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | en_US |
dc.relation.isversionof | http://dx.doi.org/10.1109/ICRA.2017.7989037 | en_US |
dc.rights | Creative Commons Attribution-Noncommercial-Share Alike | en_US |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/4.0/ | en_US |
dc.source | arXiv | en_US |
dc.title | Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Chen, Yu Fan, Miao Liu, Michael Everett, and Jonathan P. How. “Decentralized Non-Communicating Multiagent Collision Avoidance with Deep Reinforcement Learning.” 2017 IEEE International Conference on Robotics and Automation (ICRA), May 2017, Singapore, Singapore, 2017. | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Mechanical Engineering | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Laboratory for Information and Decision Systems | en_US |
dc.contributor.mitauthor | Chen, Yu Fan | |
dc.contributor.mitauthor | Liu, Miao | |
dc.contributor.mitauthor | Everett, Michael F | |
dc.contributor.mitauthor | How, Jonathan P | |
dc.relation.journal | 2017 IEEE International Conference on Robotics and Automation (ICRA) | en_US |
dc.eprint.version | Original manuscript | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2018-03-21T16:59:10Z | |
dspace.orderedauthors | Chen, Yu Fan; Liu, Miao; Everett, Michael; How, Jonathan P. | en_US |
dspace.embargo.terms | N | en_US |
dc.identifier.orcid | https://orcid.org/0000-0003-3756-3256 | |
dc.identifier.orcid | https://orcid.org/0000-0002-1648-8325 | |
dc.identifier.orcid | https://orcid.org/0000-0001-9377-6745 | |
dc.identifier.orcid | https://orcid.org/0000-0001-8576-1930 | |
mit.license | OPEN_ACCESS_POLICY | en_US |