AMDFNet: Adaptive multi-level deformable fusion network for RGB-D saliency detection

Fei Li, Jiangbin Zheng, Yuan fang Zhang, Nian Liu, Wenjing Jia

科研成果: 期刊稿件文章同行评审

7 引用 (Scopus)

摘要

Effective exploration of useful contextual information in multi-modal images is an essential task in salient object detection. Nevertheless, the existing methods based on the early-fusion or the late-fusion schemes cannot address this problem as they are unable to effectively resolve the distribution gap and information loss. In this paper, we propose an adaptive multi-level deformable fusion network (AMDFNet) to exploit the cross-modality information. We use a cross-modality deformable convolution module to dynamically adjust the boundaries of salient objects by exploring the extra input from another modality. This enables incorporating the existing features and propagating more contexts so as to strengthen the model's ability to perceiving scenes. To accurately refine the predicted maps, a multi-scaled feature refinement module is proposed to enhance the intermediate features with multi-level prediction in the decoder part. Furthermore, we introduce a selective cross-modality attention module in the fusion process to exploit the attention mechanism. This module captures dense long-range cross-modality dependencies from a multi-modal hierarchical feature's perspective. This strategy enables the network to select more informative details and suppress the contamination caused by the negative depth maps. Experimental results on eight benchmark datasets demonstrate the effectiveness of the components in our proposed model, as well as the overall saliency model.

源语言英语
页(从-至)141-156
页数16
期刊Neurocomputing
465
DOI
出版状态已出版 - 20 11月 2021

指纹

探究 'AMDFNet: Adaptive multi-level deformable fusion network for RGB-D saliency detection' 的科研主题。它们共同构成独一无二的指纹。

引用此