Monocular omnidirectional visual odometry (OVO) systems leverage 360-degree cameras to overcome field-of-view limitations of perspective VO systems. However, existing methods, reliant on handcrafted features or photometric objectives, often lack robustness in challenging scenarios, such as aggressive motion and varying illumination. To address this, we present 360DVO, the first deep learning-based OVO framework. Our approach introduces a distortion-aware spherical feature extractor (DAS-Feat) that adaptively learns distortion-resistant features from 360-degree images. These sparse feature patches are then used to establish constraints for effective pose estimation within a novel omnidirectional differentiable bundle adjustment (ODBA) module. To facilitate evaluation in realistic settings, we also contribute a new real-world OVO benchmark. Extensive experiments on this benchmark and public synthetic datasets (TartanAir V2 and 360VO) demonstrate that 360DVO surpasses state-of-the-art baselines (including 360VO and OpenVSLAM), improving robustness by 50% and accuracy by 37.5%. Homepage: https://chris1004336379.github.io/360DVO-homepage
翻译:单目全向视觉里程计(OVO)系统利用360度相机来克服透视VO系统的视野限制。然而,现有方法依赖于手工特征或光度目标,在挑战性场景(如剧烈运动和变化光照)中通常缺乏鲁棒性。为解决此问题,我们提出了360DVO,首个基于深度学习的OVO框架。我们的方法引入了一个感知畸变的球面特征提取器(DAS-Feat),它能自适应地从360度图像中学习抗畸变特征。这些稀疏特征块随后被用于在一个新颖的全向可微光束法平差(ODBA)模块中建立有效位姿估计的约束。为了促进在真实场景中的评估,我们还贡献了一个新的真实世界OVO基准。在此基准和公开合成数据集(TartanAir V2 和 360VO)上进行的大量实验表明,360DVO超越了最先进的基线方法(包括360VO和OpenVSLAM),将鲁棒性提高了50%,精度提高了37.5%。主页:https://chris1004336379.github.io/360DVO-homepage