Recurrent Convolutional Neural Networks For Object-Class Segmentation Of Rgb-D Video

2015 International Joint Conference on Neural Networks (IJCNN)(2015)

引用 30|浏览55
暂无评分
摘要
Object-class segmentation is a computer vision task which requires labeling each pixel of an image with the class of the object it belongs to. Deep convolutional neural networks (DNN) are able to learn and exploit local spatial correlations required for this task. They are, however, restricted by their small, fixed-sized filters, which limits their ability to learn long-range dependencies. Recurrent Neural Networks (RNN), on the other hand, do not suffer from this restriction. Their iterative interpretation allows them to model long-range dependencies by propagating activity. This property might be especially useful when labeling video sequences, where both spatial and temporal long-range dependencies occur. In this work, we propose novel RNN architectures for object-class segmentation. We investigate three ways to consider past and future context in the prediction process by comparing networks that process the frames one by one with networks that have access to the whole sequence. We evaluate our models on the challenging NYU Depth v2 dataset for object-class segmentation and obtain competitive results.
更多
查看译文
关键词
recurrent convolutional neural networks,RGB-D video,object-class segmentation,computer vision task,image pixel labeling,deep convolutional neural networks,DNN,local spatial correlations,fixed-sized filters,video sequences labeling,spatial long-range dependencies,temporal long-range dependencies,RNN architectures,prediction process,NYU Depth v2 dataset
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要