Abstract
This paper presents a multimodal indoor odometry dataset, OdomBeyondVision, featuring multiple sensors across the different spectrum and collected with different mobile platforms. Not only does OdomBeyondVision contain the traditional navigation sensors, sensors such as IMUs, mechanical LiDAR, RGBD camera, it also includes several emerging sensors such as the single-chip mmWave radar, LWIR thermal camera and solid-state LiDAR. With the above sensors on UAV, UGV and handheld platforms, we respectively recorded the multimodal odometry data and their movement trajectories in various indoor scenes and different illumination conditions. We release the exemplar radar, radar-inertial and thermal-inertial odometry implementations to demonstrate their results for future works to compare against and improve upon. The full dataset including toolkit and documentation is publicly available at: https://github.com/MAPS-Lab/OdomBeyondVision.
Original language | English |
---|---|
Title of host publication | IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2022 |
Publisher | IEEE |
Pages | 3845-3850 |
Number of pages | 6 |
ISBN (Electronic) | 9781665479271 |
DOIs | |
Publication status | Published - 23 Oct 2022 |
Event | 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2022 - Kyoto, Japan Duration: 23 Oct 2022 → 27 Oct 2022 |
Publication series
Name | IEEE International Conference on Intelligent Robots and Systems |
---|---|
Volume | 2022-October |
ISSN (Print) | 2153-0858 |
ISSN (Electronic) | 2153-0866 |
Conference
Conference | 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2022 |
---|---|
Country/Territory | Japan |
City | Kyoto |
Period | 23/10/22 → 27/10/22 |
Bibliographical note
Funding Information:This work was partially supported by Amazon Web Services via the Oxford-Singapore Human-Machine Collaboration Programme and EPSRC ACE-OPS (EP/S030832/1)