Fusing visual odometry and depth completion
Author(s)
Venturelli Cavalheiro, Guilherme.
Download1121277435-MIT.pdf (4.908Mb)
Other Contributors
Massachusetts Institute of Technology. Department of Aeronautics and Astronautics.
Advisor
Sertac Karaman.
Terms of use
Metadata
Show full item recordAbstract
Recent advances in technology indicate that autonomous vehicles and self-driving cats in particular may become commonplace in the near future. This thesis contributes to that scenario by studying the problem of depth perception based on sequences of camera images. We start by presenting a sensor fusion framework that achieves state-of-the-art performance when completing depth from sparse LiDAR measurements and a camera. Then, we study how the system performs under a variety of modifications of the sparse input until we ultimately replace LiDAR measurements with triangulations from a typical sparse visual odometry pipeline. We are then able to achieve a small improvement over the single image baseline and chart guidelines to assist in designing a system with even more substantial gains.
Description
Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2019 Cataloged from PDF version of thesis. Includes bibliographical references (pages 57-62).
Date issued
2019Department
Massachusetts Institute of Technology. Department of Aeronautics and AstronauticsPublisher
Massachusetts Institute of Technology
Keywords
Aeronautics and Astronautics.