Drift-free humanoid state estimation fusing kinematic, inertial and LIDAR sensing
Author(s)
Fallon, Maurice; Antone, Matthew; Roy, Nicholas; Teller, Seth
DownloadRoy_Drift-free.pdf (3.954Mb)
OPEN_ACCESS_POLICY
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
This paper describes an algorithm for the probabilistic fusion of sensor data from a variety of modalities (inertial, kinematic and LIDAR) to produce a single consistent position estimate for a walking humanoid. Of specific interest is our approach for continuous LIDAR-based localization which
maintains reliable drift-free alignment to a prior map using a Gaussian Particle Filter. This module can be bootstrapped by constructing the map on-the-fly and performs robustly in a variety of challenging field situations. We also discuss a two-tier estimation hierarchy which preserves registration to this map and other objects in the robot’s vicinity while also contributing to direct low-level control of a Boston Dynamics Atlas robot. Extensive experimental demonstrations illustrate how the approach can enable the humanoid to walk over uneven terrain without stopping (for tens of minutes), which would otherwise not be possible. We characterize the performance of the estimator for each sensor modality and discuss the
computational requirements.
Date issued
2014-11Department
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory; Massachusetts Institute of Technology. Department of Aeronautics and AstronauticsJournal
IEEE-RAS International Conference on Humanoid Robots, 2014.
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
Citation
Fallon, Maurice F. et al. “Drift-Free Humanoid State Estimation Fusing Kinematic, Inertial and LIDAR Sensing.” IEEE, 2014. 112–119.
Version: Author's final manuscript
ISBN
978-1-4799-7174-9