TY - JOUR
T1 - Scene-Embedded Generative Adversarial Networks for Semi-Supervised SAR-to-Optical Image Translation
AU - Guo, Zhe
AU - Luo, Rui
AU - Cai, Qinglin
AU - Liu, Jiayi
AU - Zhang, Zhibo
AU - Mei, Shaohui
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - SAR-to-optical image translation (S2OIT) improves the interpretability of SAR images, providing a clearer visual insight that can significantly enhance remote sensing applications. Compared to supervised S2OIT methods that are limited by the paired dataset, unsupervised methods have shown more advantages in practical applications. However, the existing unsupervised S2OIT approaches, designed for unpaired datasets, often struggle to generalize well to scenes that are significantly different from the training data, potentially leading to mistranslations in diverse scenarios. To address the above issues, we propose a scene-embedded generative adversarial network for semi-supervised S2OIT called ScE-GAN, which utilizes the scene category labels in addition to unpaired image dataset, thus effectively improving the robustness of S2OIT under different scenes without increasing complex network structure and learning cost. In particular, a scene information fusion generator (SIFG) is proposed to learn the relationship between the image and the scene directly through scene category guidance and multihead attention, enhancing its ability to adapt to scene changes. Moreover, a scene-assisted discriminator (SAD) is presented cooperating with the generator to ensure both image authenticity and scene accuracy. Extensive experiments on two challenging datasets SEN1-2 and QXS-SAROPT demonstrate that our method outperforms the state-of-the-art methods in both objective and subjective evaluations. Our code and more details are available at https://github.com/lr-dddd/ScE-GAN.
AB - SAR-to-optical image translation (S2OIT) improves the interpretability of SAR images, providing a clearer visual insight that can significantly enhance remote sensing applications. Compared to supervised S2OIT methods that are limited by the paired dataset, unsupervised methods have shown more advantages in practical applications. However, the existing unsupervised S2OIT approaches, designed for unpaired datasets, often struggle to generalize well to scenes that are significantly different from the training data, potentially leading to mistranslations in diverse scenarios. To address the above issues, we propose a scene-embedded generative adversarial network for semi-supervised S2OIT called ScE-GAN, which utilizes the scene category labels in addition to unpaired image dataset, thus effectively improving the robustness of S2OIT under different scenes without increasing complex network structure and learning cost. In particular, a scene information fusion generator (SIFG) is proposed to learn the relationship between the image and the scene directly through scene category guidance and multihead attention, enhancing its ability to adapt to scene changes. Moreover, a scene-assisted discriminator (SAD) is presented cooperating with the generator to ensure both image authenticity and scene accuracy. Extensive experiments on two challenging datasets SEN1-2 and QXS-SAROPT demonstrate that our method outperforms the state-of-the-art methods in both objective and subjective evaluations. Our code and more details are available at https://github.com/lr-dddd/ScE-GAN.
KW - Scene assist
KW - scene information fusion
KW - synthetic aperture radar (SAR)-to-optical image translation (S2OIT)
UR - http://www.scopus.com/inward/record.url?scp=85205899315&partnerID=8YFLogxK
U2 - 10.1109/LGRS.2024.3471553
DO - 10.1109/LGRS.2024.3471553
M3 - 文章
AN - SCOPUS:85205899315
SN - 1545-598X
VL - 21
JO - IEEE Geoscience and Remote Sensing Letters
JF - IEEE Geoscience and Remote Sensing Letters
M1 - 4018005
ER -