TY - JOUR
T1 - Bridging optical and SAR satellite image time series via contrastive feature extraction for crop classification
AU - Yuan, Yuan
AU - Lin, Lei
AU - Zhou, Zeng Guang
AU - Jiang, Houjun
AU - Liu, Qingshan
N1 - Publisher Copyright:
© 2022 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)
PY - 2023/1
Y1 - 2023/1
N2 - Precise crop mapping is crucial for guiding agricultural production, forecasting crop yield, and ensuring food security. Integrating optical and synthetic aperture radar (SAR) satellite image time series (SITS) for crop classification is an essential and challenging task in remote sensing. Previously published studies generally employ a dual-branch network to learn optical and SAR features independently, while ignoring the complementarity and correlation between the two modalities. In this article, we propose a novel method to learn optical and SAR features for crop classification through cross-modal contrastive learning. Specifically, we develop an updated dual-branch network with partial weight-sharing of the two branches to reduce model complexity. Furthermore, we enforce the network to map features of different modalities from the same class to nearby locations in a latent space, while samples from distinct classes are far apart, thereby learning discriminative and modality-invariant features. We conducted a comprehensive evaluation of the proposed method on a large-scale crop classification dataset. Experimental results show that our method consistently outperforms traditional supervised learning approaches, no matter the training samples are adequate or not. Our findings demonstrate that unifying the representations of optical and SAR image time series enables the network to learn more competitive features and suppress inference noise.
AB - Precise crop mapping is crucial for guiding agricultural production, forecasting crop yield, and ensuring food security. Integrating optical and synthetic aperture radar (SAR) satellite image time series (SITS) for crop classification is an essential and challenging task in remote sensing. Previously published studies generally employ a dual-branch network to learn optical and SAR features independently, while ignoring the complementarity and correlation between the two modalities. In this article, we propose a novel method to learn optical and SAR features for crop classification through cross-modal contrastive learning. Specifically, we develop an updated dual-branch network with partial weight-sharing of the two branches to reduce model complexity. Furthermore, we enforce the network to map features of different modalities from the same class to nearby locations in a latent space, while samples from distinct classes are far apart, thereby learning discriminative and modality-invariant features. We conducted a comprehensive evaluation of the proposed method on a large-scale crop classification dataset. Experimental results show that our method consistently outperforms traditional supervised learning approaches, no matter the training samples are adequate or not. Our findings demonstrate that unifying the representations of optical and SAR image time series enables the network to learn more competitive features and suppress inference noise.
KW - Contrastive learning
KW - Crop classification
KW - Feature extraction
KW - Satellite image time series (SITS)
KW - Synthetic aperture radar (SAR)
UR - http://www.scopus.com/inward/record.url?scp=85144031329&partnerID=8YFLogxK
U2 - 10.1016/j.isprsjprs.2022.11.020
DO - 10.1016/j.isprsjprs.2022.11.020
M3 - 文章
AN - SCOPUS:85144031329
SN - 0924-2716
VL - 195
SP - 222
EP - 232
JO - ISPRS Journal of Photogrammetry and Remote Sensing
JF - ISPRS Journal of Photogrammetry and Remote Sensing
ER -