dc.contributor.advisor | Nicholas Roy. | en_US |
dc.contributor.author | Greene, W. Nicholas(William Nicholas) | en_US |
dc.contributor.other | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics. | en_US |
dc.date.accessioned | 2021-05-24T20:22:42Z | |
dc.date.available | 2021-05-24T20:22:42Z | |
dc.date.copyright | 2021 | en_US |
dc.date.issued | 2021 | en_US |
dc.identifier.uri | https://hdl.handle.net/1721.1/130747 | |
dc.description | Thesis: Ph. D., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, February, 2021 | en_US |
dc.description | Cataloged from the official PDF of thesis. | en_US |
dc.description | Includes bibliographical references (pages 135-151). | en_US |
dc.description.abstract | Monocular cameras are powerful sensors for a variety of computer vision tasks since they are small, inexpensive, and provide dense perceptual information about the surrounding environment. Efficiently estimating the pose of a moving monocular camera and the 3D structure of the observed scene from the images alone is a fundamental problem in computer vision commonly referred to as monocular simultaneous localization and mapping (SLAM). Given the importance of egomotion estimation and environmental mapping to many applications in robotics and augmented reality, the last twenty years have seen dramatic advances in the state of the art in monocular SLAM. Despite the rapid progress, however, several limitations remain that prevent monocular SLAM systems from transitioning out of the research laboratory and into large, uncontrolled environments on small, resource-constrained computing platforms. This thesis presents research that attempts to address existing problems in monocular SLAM by leveraging different sources of prior information along with targeted applications of machine learning. First, we exploit the piecewise planar structure common in many environments in order to represent the scene using compact triangular meshes that will allow for faster reconstruction and regularization. Second, we leverage the semantic information encoded in large datasets of images to constrain the unobservable scale of motion of the monocular solution to the true, metric scale without additional sensors. Lastly, we compensate for known viewpoint changes when associating pixels between images in order to allow for robust, learning-based depth estimation across disparate views. | en_US |
dc.description.statementofresponsibility | by W. Nicholas Greene. | en_US |
dc.format.extent | 151 pages | en_US |
dc.language.iso | eng | en_US |
dc.publisher | Massachusetts Institute of Technology | en_US |
dc.rights | MIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided. | en_US |
dc.rights.uri | http://dspace.mit.edu/handle/1721.1/7582 | en_US |
dc.subject | Aeronautics and Astronautics. | en_US |
dc.title | Leveraging prior information for real-time monocular simultaneous localization and mapping | en_US |
dc.type | Thesis | en_US |
dc.description.degree | Ph. D. | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics | en_US |
dc.identifier.oclc | 1251896691 | en_US |
dc.description.collection | Ph.D. Massachusetts Institute of Technology, Department of Aeronautics and Astronautics | en_US |
dspace.imported | 2021-05-24T20:22:42Z | en_US |
mit.thesis.degree | Doctoral | en_US |
mit.thesis.department | Aero | en_US |