A Novel Cross-Attention-Based Pedestrian Visual–Inertial Odometry With Analyses Demonstrating Challenges in Dense Optical Flow

Ilari Pajula,Niclas Joswig,Aiden Morrison, Nadia Sokolova,Laura Ruotsalainen

IEEE Journal of Indoor and Seamless Positioning and Navigation(2024)

引用 0|浏览0
暂无评分
摘要
Visual–inertial odometry (VIO), the fusion of visual and inertial sensor data, has been shown to be functional for navigation in global-navigation-satellite-system-denied environments. Recently, dense-optical-flow-based end-to-end trained deep learning VIO models have gained superior performance in outdoor navigation. In this article, we introduced a novel visual–inertial sensor fusion approach based on vision transformer architecture with a cross-attention mechanism, specifically designed to better integrate potentially poor-quality optical flow features with inertial data. Although optical-flow-based VIO models have obtained superior performance in outdoor vehicle navigation, both in accuracy and ease of calibration, we have shown how their suitability for indoor pedestrian navigation is still far from existing feature-matching-based methods. We compare the performance of traditional VIO models against deep-learning-based VIO models on the KITTI benchmark dataset and our custom pedestrian navigation dataset. We show how end-to-end trained VIO models using optical flow were significantly outperformed by simpler visual odometry models utilizing feature matching. Our findings indicate that due to the robustness against occlusion and camera shake, feature matching is better suited for indoor pedestrian navigation, whereas dense optical flow remains viable for vehicular data. Therefore, the most feasible way forward will be the integration of our novel model with feature-based visual data encoding.
更多
查看译文
关键词
Deep learning,sensor fusion,visual–inertial odometry (VIO)
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要