Exploiting textual queries for dynamically visual disambiguation

Zeren Sun, Yazhou Yao, Jimin Xiao, Lei Zhang, Jian Zhang, Zhenmin Tang

科研成果: 期刊稿件文章同行评审

13 引用 (Scopus)

摘要

Due to the high cost of manual annotation, learning directly from the web has attracted broad attention. One issue that limits the performance of current webly supervised models is the problem of visual polysemy. In this work, we present a novel framework that resolves visual polysemy by dynamically matching candidate text queries with retrieved images. Specifically, our proposed framework includes three major steps: we first discover and then dynamically select the text queries according to the keyword-based image search results, we employ the proposed saliency-guided deep multi-instance learning (MIL) network to remove outliers and learn classification models for visual disambiguation. Compared to existing methods, our proposed approach can figure out the right visual senses, adapt to dynamic changes in the search results, remove outliers, and jointly learn the classification models. Extensive experiments and ablation studies on CMU-Poly-30 and MIT-ISD datasets demonstrate the effectiveness of our proposed approach.

源语言英语
文章编号107620
期刊Pattern Recognition
110
DOI
出版状态已出版 - 2月 2021

指纹

探究 'Exploiting textual queries for dynamically visual disambiguation' 的科研主题。它们共同构成独一无二的指纹。

引用此