Multi-modal 3D Human Pose Estimation with 2D Weak Supervision in Autonomous Driving

IEEE Conference on Computer Vision and Pattern Recognition(2021)

引用 25|浏览102
暂无评分
摘要
3D human pose estimation (HPE) in autonomous vehicles (AV) differs from other use cases in many factors, including the 3D resolution and range of data, absence of dense depth maps, failure modes for LiDAR, relative location between the camera and LiDAR, and a high bar for estimation accuracy. Data collected for other use cases (such as virtual reality, gaming, and animation) may therefore not be usable for AV applications. This necessitates the collection and annotation of a large amount of 3D data for HPE in AV, which is time-consuming and expensive.In this paper, we propose one of the first approaches to alleviate this problem in the AV setting. Specifically, we propose a multi-modal approach which uses 2D labels on RGB images as weak supervision to perform 3D HPE. The proposed multi-modal architecture incorporates LiDAR and camera inputs with an auxiliary segmentation branch. On the Waymo Open Dataset [27], our approach achieves a ∼ 22% relative improvement over camera-only 2D HPE baseline, and ∼ 6% improvement over LiDAR-only model. Finally, careful ablation studies and parts based analysis illustrate the advantages of each of our contributions.
更多
查看译文
关键词
multimodal approach,HPE,multimodal architecture,camera inputs,22% relative improvement,LiDAR-only model,multimodal 3D human pose estimation,autonomous driving,autonomous vehicles,dense depth maps,failure modes,relative location,high bar,estimation accuracy,AV applications,AV setting
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要