Double cross-modality progressively guided network for RGB-D salient object detection

Image and Vision Computing(2022)

引用 4|浏览26
暂无评分
摘要
With the advent of depth sensors, the use of RGB and depth(D) information for salient object detection (SOD) has been explored extensively in recent years. However, the depth quality from the different scenes usually varies, leading to fusing and achieving complementary between RGB and low-quality depth is stilla challenging prob-lem. In this paper, we first design a Double Dilated Merge Module (DDMM) to extract comprehensive and ben-eficial high-level cross-modality features and explore further global context information at multi-scales to obtain a coarse saliency map. Then, we propose a Cross-Modality Enhance Module (CMEM) to enhance cross-modality features compatibility and fuse them with the previous predicted coarse saliency map to generate a more accu-rate saliency map. Furthermore, we introduce a Distribution-Region Combination Loss (DRcom Loss) to optimize our proposed Double Cross-Modality Progressively Guided Network (DCPGNet) in a coarse-to-fine manner. DCPGNet achieves satisfactory performance on five public benchmarks compared with recent state-of-the-art algorithms. (c) 2021 Elsevier B.V. All rights reserved.
更多
查看译文
关键词
RGB-D,Salient object detection,Cross-modality,Attention mechanism,Integration
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要