Show simple item record

dc.contributor.authorArkin, Jacob
dc.contributor.authorPark, Daehyung
dc.contributor.authorRoy, Subhro
dc.contributor.authorWalter, Matthew R
dc.contributor.authorRoy, Nicholas
dc.contributor.authorHoward, Thomas M
dc.contributor.authorPaul, Rohan
dc.date.accessioned2021-05-12T13:36:09Z
dc.date.available2021-05-12T13:36:09Z
dc.date.issued2020-06
dc.identifier.issn0278-3649
dc.identifier.issn1741-3176
dc.identifier.urihttps://hdl.handle.net/1721.1/130573
dc.description.abstractThe goal of this article is to enable robots to perform robust task execution following human instructions in partially observable environments. A robot’s ability to interpret and execute commands is fundamentally tied to its semantic world knowledge. Commonly, robots use exteroceptive sensors, such as cameras or LiDAR, to detect entities in the workspace and infer their visual properties and spatial relationships. However, semantic world properties are often visually imperceptible. We posit the use of non-exteroceptive modalities including physical proprioception, factual descriptions, and domain knowledge as mechanisms for inferring semantic properties of objects. We introduce a probabilistic model that fuses linguistic knowledge with visual and haptic observations into a cumulative belief over latent world attributes to infer the meaning of instructions and execute the instructed tasks in a manner robust to erroneous, noisy, or contradictory evidence. In addition, we provide a method that allows the robot to communicate knowledge dissonance back to the human as a means of correcting errors in the operator’s world model. Finally, we propose an efficient framework that anticipates possible linguistic interactions and infers the associated groundings for the current world state, thereby bootstrapping both language understanding and generation. We present experiments on manipulators for tasks that require inference over partially observed semantic properties, and evaluate our framework’s ability to exploit expressed information and knowledge bases to facilitate convergence, and generate statements to correct declared facts that were observed to be inconsistent with the robot’s estimate of object properties.en_US
dc.description.sponsorshipARO (Grant W911NF-15-1-0402)en_US
dc.description.sponsorshipToyota Research Institute (Award LPC000765- SR)en_US
dc.publisherSAGE Publicationsen_US
dc.relation.isversionofhttp://dx.doi.org/10.1177/0278364920917755en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceProf. Royen_US
dc.titleMultimodal estimation and communication of latent semantic knowledge for robust execution of robot instructionsen_US
dc.typeArticleen_US
dc.identifier.citationArkin, Jacob et al. "Multimodal estimation and communication of latent semantic knowledge for robust execution of robot instructions." International Journal of Robotics Research 39, 10-11 (September 2020): 1279-1304. © 2020 SAGE Publicationsen_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.contributor.approverRoy, Nicholasen_US
dc.relation.journalInternational Journal of Robotics Researchen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dspace.date.submission2021-05-07T19:47:39Z
mit.journal.volume39en_US
mit.journal.issue10-11en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record