TY - GEN
T1 - SAR and Optical Satellite Image Matching Incorporating Category-Supervised Semantic Features
AU - Zhou, Jia
AU - Zhao, Chunhui
AU - Lyu, Yang
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The matching of SAR and optical satellite images is a crucial method for addressing the challenge of geolocalization in scenarios where real-time access to optical images is constrained (e.g., darkness, rain, fog, glare, etc.). To enhance the generalization capability of SAR-optical image matching, particularly under varying viewpoints (rotation), this paper presents an augmented Superpoint network architecture. This architecture incorporates semantic segmentation feature representation to facilitate a priori semantic description and to integrate it with the original network's feature descriptors. This integration is achieved by introducing self-attention and cross-attention layers, which enhance the feature vectors. The experiments demonstrate that, compared to the original architecture, the model reinforces the relationship for commonality matching and significantly increases the number of correctly matched points. Compared with other deep learning methods, including CMMNet and LoFTR, the matching scheme presented in this paper is less time-consuming and more robust with anti-rotation capability.
AB - The matching of SAR and optical satellite images is a crucial method for addressing the challenge of geolocalization in scenarios where real-time access to optical images is constrained (e.g., darkness, rain, fog, glare, etc.). To enhance the generalization capability of SAR-optical image matching, particularly under varying viewpoints (rotation), this paper presents an augmented Superpoint network architecture. This architecture incorporates semantic segmentation feature representation to facilitate a priori semantic description and to integrate it with the original network's feature descriptors. This integration is achieved by introducing self-attention and cross-attention layers, which enhance the feature vectors. The experiments demonstrate that, compared to the original architecture, the model reinforces the relationship for commonality matching and significantly increases the number of correctly matched points. Compared with other deep learning methods, including CMMNet and LoFTR, the matching scheme presented in this paper is less time-consuming and more robust with anti-rotation capability.
KW - Cross modality remote sensing image matching
KW - self-cross attention mechanism
KW - semantic matching
KW - synthetic aperture radar (SAR)-optical image
UR - http://www.scopus.com/inward/record.url?scp=105002217481&partnerID=8YFLogxK
U2 - 10.1109/SWC62898.2024.00299
DO - 10.1109/SWC62898.2024.00299
M3 - 会议稿件
AN - SCOPUS:105002217481
T3 - Proceedings - 2024 IEEE Smart World Congress, SWC 2024 - 2024 IEEE Ubiquitous Intelligence and Computing, Autonomous and Trusted Computing, Digital Twin, Metaverse, Privacy Computing and Data Security, Scalable Computing and Communications
SP - 1949
EP - 1954
BT - Proceedings - 2024 IEEE Smart World Congress, SWC 2024 - 2024 IEEE Ubiquitous Intelligence and Computing, Autonomous and Trusted Computing, Digital Twin, Metaverse, Privacy Computing and Data Security, Scalable Computing and Communications
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 10th IEEE Smart World Congress, SWC 2024
Y2 - 2 December 2024 through 7 December 2024
ER -