Full Surround Monodepth From Multiple Cameras

IEEE ROBOTICS AND AUTOMATION LETTERS(2022)

引用 13|浏览53
暂无评分
摘要
Self-supervised monocular depth and ego-motion estimation is a promising approach to replace or supplement expensive depth sensors such as LiDAR for robotics applications like autonomous driving. However, most research in this area focuses on a single monocular camera or stereo pairs that cover only a fraction of the scene around the vehicle. In this work, we extend monocular self-supervised depth and ego-motion estimation to large-baseline multi-camera rigs. Using generalized spatio-temporal contexts, pose consistency constraints, and carefully designed photometric loss masking, we learn a single network generating dense, consistent, and scale-aware point clouds that cover the same full surround 360 degrees field of view as a typical LiDAR scanner. We also propose a new scale-consistent evaluation metric more suitable to multicamera settings. Experiments on two challenging benchmarks illustrate the benefits of our approach over strong baselines.
更多
查看译文
关键词
Computer vision, machine learning, autonomous automobiles
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要