Enhanced Radar Perception via Multi-Task Learning: Towards Refined Data for Sensor Fusion Applications
arxiv(2024)
摘要
Radar and camera fusion yields robustness in perception tasks by leveraging
the strength of both sensors. The typical extracted radar point cloud is 2D
without height information due to insufficient antennas along the elevation
axis, which challenges the network performance. This work introduces a
learning-based approach to infer the height of radar points associated with 3D
objects. A novel robust regression loss is introduced to address the sparse
target challenge. In addition, a multi-task training strategy is employed,
emphasizing important features. The average radar absolute height error
decreases from 1.69 to 0.25 meters compared to the state-of-the-art height
extension method. The estimated target height values are used to preprocess and
enrich radar data for downstream perception tasks. Integrating this refined
radar information further enhances the performance of existing radar camera
fusion models for object detection and depth estimation tasks.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要