Like Humans to Few-Shot Learning Through Knowledge Permeation of Visual and Language

Yuyu Jia, Qing Zhou, Junyu Gao, Qiang Li, Qi Wang

Research output: Contribution to journalArticlepeer-review

Abstract

Few-shot learning aims to generalize the recognizer from seen categories to an entirely novel scenario. With only a few support samples, several advanced methods initially introduce class names as prior knowledge for identifying novel classes. However, obstacles still impede achieving a comprehensive understanding of how to harness the mutual advantages of visual and textual knowledge. In this paper, we set out to fill this gap via a coherent Bidirectional Knowledge Permeation strategy called BiKop, which is grounded in human intuition: a class name description offers a more general representation, whereas an image captures the specificity of individuals. BiKop primarily establishes a hierarchical joint general-specific representation through bidirectional knowledge permeation. On the other hand, considering the bias of joint representation towards the base set, we disentangle base-class-relevant semantics during training, thereby alleviating the suppression of potential novel-class-relevant information. Experiments on four challenging benchmarks demonstrate the remarkable superiority of BiKop, particularly outperforming previous methods by a substantial margin in the 1-shot setting (improving the accuracy by 7.58% on miniImageNet).

Original languageEnglish
Pages (from-to)7905-7916
Number of pages12
JournalIEEE Transactions on Multimedia
Volume27
DOIs
StatePublished - 2025

Keywords

  • Few-shot learning
  • class-relevant information
  • knowledge disparity

Fingerprint

Dive into the research topics of 'Like Humans to Few-Shot Learning Through Knowledge Permeation of Visual and Language'. Together they form a unique fingerprint.

Cite this