Single-Stream Extractor Network with Contrastive Pre-Training for Remote-Sensing Change Captioning

科研成果: 期刊稿件文章同行评审

6 引用 (Scopus)

摘要

Remote-sensing (RS) image change captioning (RSICC) is a visual semantic understanding task that has received increasing attention. The change captioning methods are required to understand the visual information of the images and capture the most significant difference between them, then describe it in natural language. Most existing methods mainly focus on improving the difference feature encoder or language decoder, while ignoring the visual feature extractor. The current feature extractors suffer from several issues, including: 1) domain gap between pre-training on single-temporal natural images and downstream bitemporal RS task; 2) limited difference feature modeling in the implicit single-stream network; and 3) high computational costs caused by extracting features for each temporal phase image under the dual-stream extractor. To address these issues, we propose a single-stream extractor network (SEN). It consists of a single-stream extractor pre-trained on bitemporal RS images using contrastive learning to mitigate the domain gap and high computational cost. Additionally, to improve feature modeling for difference information, we propose a shallow feature embedding (SFE) module and a cross-attention guided difference (CAGD) module, which enhance the representation of temporal features and extract the difference features explicitly. Extensive experiments and visualizations demonstrate the effectiveness and advanced performance of SEN. The code and model weights are available at https://github.com/mrazhou/SEN.

源语言英语
文章编号5624514
页(从-至)1-14
页数14
期刊IEEE Transactions on Geoscience and Remote Sensing
62
DOI
出版状态已出版 - 2024

指纹

探究 'Single-Stream Extractor Network with Contrastive Pre-Training for Remote-Sensing Change Captioning' 的科研主题。它们共同构成独一无二的指纹。

引用此