TY - JOUR
T1 - Smooth fusion of multi-spectral images via total variation minimization for traffic scene semantic segmentation
AU - Li, Ying
AU - Fang, Aiqing
AU - Guo, Yangming
AU - Sun, Wei
AU - Yang, Xiaobao
AU - Wang, Xiaodong
N1 - Publisher Copyright:
© 2023 Elsevier Ltd
PY - 2024/4
Y1 - 2024/4
N2 - Achieving precise semantic segmentation for traffic scenes relies on adopting multi-spectral image fusion techniques to attain high-quality images. Many existing fusion solutions often aim to enhance the similarity between the input and fusion results at the pixel intensity and texture details stage. However, this can result in smoothness issues that limit semantic segmentation performance. To address these issues, we present a smooth representation learning optimization mechanism (SFLM) that conducts image fusion on two dimensions: inter- and intra-image levels. The former overcomes over- or under-smoothing problems via the mutual information maximization between the fusion result and image samples (i.e., negative and positive). The latter balances under and over-smoothing for fusion results by minimizing the total variation in pixel space and maximizing the total variation in gradient space based on contrast learning. In this way, the proposed method effectively overcomes the fusion quality issues, providing better feature representations for semantic segmentation in autonomous vehicles. Experimental results on four public datasets validate our method's effectiveness, robustness, and overall superiority.
AB - Achieving precise semantic segmentation for traffic scenes relies on adopting multi-spectral image fusion techniques to attain high-quality images. Many existing fusion solutions often aim to enhance the similarity between the input and fusion results at the pixel intensity and texture details stage. However, this can result in smoothness issues that limit semantic segmentation performance. To address these issues, we present a smooth representation learning optimization mechanism (SFLM) that conducts image fusion on two dimensions: inter- and intra-image levels. The former overcomes over- or under-smoothing problems via the mutual information maximization between the fusion result and image samples (i.e., negative and positive). The latter balances under and over-smoothing for fusion results by minimizing the total variation in pixel space and maximizing the total variation in gradient space based on contrast learning. In this way, the proposed method effectively overcomes the fusion quality issues, providing better feature representations for semantic segmentation in autonomous vehicles. Experimental results on four public datasets validate our method's effectiveness, robustness, and overall superiority.
KW - Image fusion and segmentation
KW - Neural network
KW - Self-supervised learning
KW - Total variation theory
UR - http://www.scopus.com/inward/record.url?scp=85180555219&partnerID=8YFLogxK
U2 - 10.1016/j.engappai.2023.107741
DO - 10.1016/j.engappai.2023.107741
M3 - 文章
AN - SCOPUS:85180555219
SN - 0952-1976
VL - 130
JO - Engineering Applications of Artificial Intelligence
JF - Engineering Applications of Artificial Intelligence
M1 - 107741
ER -