TY - GEN
T1 - MFCS-Depth
T2 - 14th IEEE International Conference on Signal Processing, Communications and Computing, ICSPCC 2024
AU - Cheng, Zeyu
AU - Zhang, Yi
AU - Zhu, Xingxing
AU - Yu, Yang
AU - Song, Zhe
AU - Tang, Chengkai
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Self-supervised monocular depth estimation plays an extremely important role in fields such as autonomous driving and intelligent robot navigation. However, general monocular depth estimation models require massive computing resources, which seriously hinders their deployment on mobile devices, which is urgently needed in fields such as autonomous driving. To address this problem, we propose MFCS-Depth, an economical monocular depth estimation method based on multi-scale fusion and channel separation attention mechanism. We use the Transformer architecture with linear self-attention as its encoder to ensure its global modeling and economy. A high-performance and low-cost decoder has also been designed to improve the local and global reasoning of the network through multi-scale attention fusion and uses scale-wise channel separation to reduce parameters and computing costs significantly. Extensive experiments show that MFCS-Depth achieves competitive results with very few parameters on the KITTI and DDAD datasets and achieves state-of-the-art performance among methods of similar size.
AB - Self-supervised monocular depth estimation plays an extremely important role in fields such as autonomous driving and intelligent robot navigation. However, general monocular depth estimation models require massive computing resources, which seriously hinders their deployment on mobile devices, which is urgently needed in fields such as autonomous driving. To address this problem, we propose MFCS-Depth, an economical monocular depth estimation method based on multi-scale fusion and channel separation attention mechanism. We use the Transformer architecture with linear self-attention as its encoder to ensure its global modeling and economy. A high-performance and low-cost decoder has also been designed to improve the local and global reasoning of the network through multi-scale attention fusion and uses scale-wise channel separation to reduce parameters and computing costs significantly. Extensive experiments show that MFCS-Depth achieves competitive results with very few parameters on the KITTI and DDAD datasets and achieves state-of-the-art performance among methods of similar size.
KW - economical self-supervised monocular depth estimation
KW - multi-scale attention fusion
KW - scale-wise channel separation
KW - Transformer
UR - http://www.scopus.com/inward/record.url?scp=85214896925&partnerID=8YFLogxK
U2 - 10.1109/ICSPCC62635.2024.10770511
DO - 10.1109/ICSPCC62635.2024.10770511
M3 - 会议稿件
AN - SCOPUS:85214896925
T3 - 2024 IEEE International Conference on Signal Processing, Communications and Computing, ICSPCC 2024
BT - 2024 IEEE International Conference on Signal Processing, Communications and Computing, ICSPCC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 19 August 2024 through 22 August 2024
ER -