Enhancing Multimodal Fusion with only Unimodal Data

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

With recent advances in remote sensing technology, a wealth of multimodal data is available for applications. However, considering the domain differences between multimodal data and the alignment challenges in practical applications, it becomes important and challenging to integrate these data effectively. In this paper, we propose a multimodal prototype representation fusion network (MPRFN) for SAR and optical image fusion segmentation. Specifically, a more robust multimodal feature representation is provided by constructing multimodal category prototype representations that better capture the characteristics and distribution of each data. Meanwhile, a prototype-consistent semi-supervised learning method is proposed to improve the effectiveness of multimodal fusion semantic segmentation using a large number of unlabelled unimodal SAR images. Experiments on SAR and optical multimodal datasets show that the proposed method achieves state-of-the-art performance.

Original languageEnglish
Title of host publicationIGARSS 2024 - 2024 IEEE International Geoscience and Remote Sensing Symposium, Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages2962-2965
Number of pages4
ISBN (Electronic)9798350360325
DOIs
StatePublished - 2024
Event2024 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2024 - Athens, Greece
Duration: 7 Jul 202412 Jul 2024

Publication series

NameInternational Geoscience and Remote Sensing Symposium (IGARSS)

Conference

Conference2024 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2024
Country/TerritoryGreece
CityAthens
Period7/07/2412/07/24

Keywords

  • multimodal fusion
  • remote sensing
  • semantic segmentation
  • Semi-supervised learning

Fingerprint

Dive into the research topics of 'Enhancing Multimodal Fusion with only Unimodal Data'. Together they form a unique fingerprint.

Cite this