Constraining multimodal distribution for domain adaptation in stereo matching

Zhelun Shen, Zhuo Li, Chenming Wu, Zhibo Rao, Lina Liu, Yuchao Dai, Liangjun Zhang

Research output: Contribution to journalArticlepeer-review

Abstract

Recently, learning-based stereo matching methods have achieved great improvement in public benchmarks, where soft argmin and smooth L1 loss play core contributions to its success. However, in unsupervised domain adaptation scenarios, we observe that these two operations often yield multimodal disparity probability distributions in target domains, resulting in degraded generalization. In this paper, we propose a novel approach, Constrain Multi-modal Distribution (CMD), to address this issue. Specifically, we introduce uncertainty-regularized minimization and anisotropic soft argmin to encourage the network to produce predominantly unimodal disparity distributions in the target domain, thereby improving prediction accuracy. Experimentally, we apply the proposed method to multiple representative stereo-matching networks and conduct domain adaptation from synthetic data to unlabeled real-world scenes. Results consistently demonstrate improved generalization in both top-performing and domain-adaptable stereo-matching models. The code for CMD will be available at: https://github.com/gallenszl/CMD.

Original languageEnglish
Article number111727
JournalPattern Recognition
Volume167
DOIs
StatePublished - Nov 2025

Keywords

  • Domain adaptation
  • Stereo matching
  • Unimodal distribution

Fingerprint

Dive into the research topics of 'Constraining multimodal distribution for domain adaptation in stereo matching'. Together they form a unique fingerprint.

Cite this