TY - JOUR
T1 - Texture-Aware Causal Feature Extraction Network for Multimodal Remote Sensing Data Classification
AU - Xu, Zhengyi
AU - Jiang, Wen
AU - Geng, Jie
N1 - Publisher Copyright:
© 1980-2012 IEEE.
PY - 2024
Y1 - 2024
N2 - The pixel-level classification of multimodal remote sensing (RS) images plays a crucial role in the intelligent interpretation of RS data. However, existing methods that mainly focus on feature interaction and fusion fail to address the challenges posed by confounders - brought by sensor imaging bias, limiting their performance. In this article, we introduce causal inference into intelligent interpretation of RS and propose a new texture-aware causal feature extraction network (TeACFNet) for pixel-level fusion classification. Specifically, we propose a two-stage causal feature extraction (CFE) framework that helps networks learn more explicit class representations by capturing the causal relationships between multimodal heterogeneous data. In addition, to solve the problem of low-resolution land cover feature representation in RS images, we propose the refined statistical texture extraction (ReSTE) module. This module integrates the semantics of statistical textures in shallow feature maps through feature refinement, quantization, and encoding. Extensive experiments on two publicly available datasets with different modalities, including Houston2013 and Berlin datasets, demonstrate the remarkable effectiveness of our proposed method, which reaches a new state-of-the-art.
AB - The pixel-level classification of multimodal remote sensing (RS) images plays a crucial role in the intelligent interpretation of RS data. However, existing methods that mainly focus on feature interaction and fusion fail to address the challenges posed by confounders - brought by sensor imaging bias, limiting their performance. In this article, we introduce causal inference into intelligent interpretation of RS and propose a new texture-aware causal feature extraction network (TeACFNet) for pixel-level fusion classification. Specifically, we propose a two-stage causal feature extraction (CFE) framework that helps networks learn more explicit class representations by capturing the causal relationships between multimodal heterogeneous data. In addition, to solve the problem of low-resolution land cover feature representation in RS images, we propose the refined statistical texture extraction (ReSTE) module. This module integrates the semantics of statistical textures in shallow feature maps through feature refinement, quantization, and encoding. Extensive experiments on two publicly available datasets with different modalities, including Houston2013 and Berlin datasets, demonstrate the remarkable effectiveness of our proposed method, which reaches a new state-of-the-art.
KW - Causal feature extraction
KW - feature fusion
KW - image pixel-level classification
KW - multimodal remote sensing (RS)
KW - texture representation learning
UR - http://www.scopus.com/inward/record.url?scp=85186110033&partnerID=8YFLogxK
U2 - 10.1109/TGRS.2024.3368091
DO - 10.1109/TGRS.2024.3368091
M3 - 文章
AN - SCOPUS:85186110033
SN - 0196-2892
VL - 62
SP - 1
EP - 12
JO - IEEE Transactions on Geoscience and Remote Sensing
JF - IEEE Transactions on Geoscience and Remote Sensing
M1 - 5103512
ER -