Adapt Anything: Tailor Any Image Classifier across Domains And Categories Using Text-to-Image Diffusion Models

Weijie Chen, Haoyu Wang, Shicai Yang, Lei Zhang, Wei Wei, Yanning Zhang, Luojun Lin, Di Xie, Yueting Zhuang

Research output: Contribution to journalArticlepeer-review

Abstract

We study a novel problem in this manuscript, that is, if a modern text-to-image diffusion model can tailor any image classifier across domains and categories. Existing domain adaption works exploit both source and target data for domain alignment so as to transfer the knowledge from the labeled source data to the unlabeled target data. However, as the development of text-to-image diffusion models, we wonder if the high-fidelity synthetic data can serve as a surrogate of the source data in real world. In this way, we do not need to collect and annotate the source data for each image classification task in a one-for-one manner. Instead, we utilize only one off-the-shelf text-to-image model to synthesize images with labels derived from text prompts, and then leverage them as a bridge to dig out the knowledge from the task-agnostic text-to-image generator to the task-oriented image classifier via domain adaptation. Such a one-for-all adaptation paradigm allows us to adapt anything in the world using only one text-to-image generator as well as any unlabeled target data. Extensive experiments validate the feasibility of this idea, which even surprisingly surpasses the state-of-the-art domain adaptation works using the source data collected and annotated in real world.

Original languageEnglish
JournalIEEE Transactions on Big Data
DOIs
StateAccepted/In press - 2025

Keywords

  • Data Synthesis
  • Prompt Diversification
  • Text-to-Image Diffusion Models
  • Unsupervised Domain Adaptation

Fingerprint

Dive into the research topics of 'Adapt Anything: Tailor Any Image Classifier across Domains And Categories Using Text-to-Image Diffusion Models'. Together they form a unique fingerprint.

Cite this