Understanding 3D Semantic Structure around the Vehicle with Monocular Cameras

2018 IEEE Intelligent Vehicles Symposium (IV)(2018)

引用 12|浏览1
暂无评分
摘要
In this paper, we propose a method to recognize semantic and geometric structure of a traffic scene using monocular cameras. We designed Deep Neural Networks (DNNs) for semantic segmentation and depth estimation and trained them using data collected with a test vehicle, on top of which a 360degree panoramic camera system and a LIDAR are mounted. Collected images were manually annotated for semantic segmentation. Experimental results show that the trained DNNs can accurately classify each pixel and also accurately estimate depth of each pixel of images in validation data. Global average of semantic segmentation reached 96.4%, while overall accuracy of depth estimation was 88.7%. Generalization capability for both tasks was also tested with DNNs trained only with front facing camera images, resulting that semantic segmentation and depth estimation were successfully executed at slightly less accuracy. We also developed a novel interface using a head mount display, that enables us to evaluate results of estimation intuitively for checking how well the estimation of proposed DNNs is.
更多
查看译文
关键词
understanding 3d semantic structure,monocular cameras,geometric structure,Deep Neural Networks,semantic segmentation,depth estimation,test vehicle,360degree panoramic camera system,collected images,trained DNNs,LIDAR,front facing camera images
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要