TY - GEN
T1 - Depth-guided Deformable Convolutions for RGB-D Saliency Object Detection
AU - Li, Fei
AU - Zheng, Jiangbin
AU - Zhang, Yuan Fang
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Recently, RGB-D salient object detection(SOD) has attracted increasing research interests, and existing methods have achieved huge success owing to well-designed feature extraction and fusion. However, in existing methods, the depth maps cannot be utilized entirely since RGB and depth are usually concatenated together as an entirety and then feed into the backbone to extract features, which cannot achieve the spatial supervision between both modals. In this letter, we propose a Depth-guided Deformable 3D Convolution (Guided-Conv) to solve this problem. Specifically, the Guided-Conv obtains the sampling offset of the 3D convolution kernel guided by the extra depth input, enabling the convolutional layer to change the receptive field and adapt to geometric cross-modal transformations. Besides, the Guided-Conv also incorporates geometric cues into the forward propagation by producing spatially adaptive filter weights. Based on comprehensive experiments on several extensively used benchmarks, the Guided-Conv yields strong results against several state-of-the-art RGB-D SOD approaches based on four key evaluation metrics.
AB - Recently, RGB-D salient object detection(SOD) has attracted increasing research interests, and existing methods have achieved huge success owing to well-designed feature extraction and fusion. However, in existing methods, the depth maps cannot be utilized entirely since RGB and depth are usually concatenated together as an entirety and then feed into the backbone to extract features, which cannot achieve the spatial supervision between both modals. In this letter, we propose a Depth-guided Deformable 3D Convolution (Guided-Conv) to solve this problem. Specifically, the Guided-Conv obtains the sampling offset of the 3D convolution kernel guided by the extra depth input, enabling the convolutional layer to change the receptive field and adapt to geometric cross-modal transformations. Besides, the Guided-Conv also incorporates geometric cues into the forward propagation by producing spatially adaptive filter weights. Based on comprehensive experiments on several extensively used benchmarks, the Guided-Conv yields strong results against several state-of-the-art RGB-D SOD approaches based on four key evaluation metrics.
KW - 3D Convolution
KW - Generate Offset
KW - RGB-D
KW - Salient Object Detection
UR - http://www.scopus.com/inward/record.url?scp=85123792205&partnerID=8YFLogxK
U2 - 10.1109/CCISP52774.2021.9639345
DO - 10.1109/CCISP52774.2021.9639345
M3 - 会议稿件
AN - SCOPUS:85123792205
T3 - Proceedings - 2021 6th International Conference on Communication, Image and Signal Processings, CCISP 2021
SP - 234
EP - 239
BT - Proceedings - 2021 6th International Conference on Communication, Image and Signal Processings, CCISP 2021
A2 - Zhang, Jing
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 6th International Conference on Communication, Image and Signal Processings, CCISP 2021
Y2 - 20 November 2021
ER -