Single- and cross-modality near duplicate image pairs detection via spatial transformer comparing CNN

Yi Zhang, Shizhou Zhang, Ying Li, Yanning Zhang

科研成果: 期刊稿件文章同行评审

17 引用 (Scopus)

摘要

Recently, both single modality and cross modality near-duplicate image detection tasks have received wide attention in the community of pattern recognition and computer vision. Existing deep neural networks-based methods have achieved remarkable performance in this task. However, most of the methods mainly focus on the learning of each image from the image pair, thus leading to less use of the information between the near duplicate image pairs to some extent. In this paper, to make more use of the correlations between image pairs, we propose a spatial transformer comparing convolutional neural network (CNN) model to compare near-duplicate image pairs. Specifically, we firstly propose a comparing CNN framework, which is equipped with a cross-stream to fully learn the correlation information between image pairs, while considering the features of each image. Furthermore, to deal with the local deformations led by cropping, translation, scaling, and nonrigid transformations, we additionally introduce a spatial transformer comparing CNN model by incorporating a spatial transformer module to the comparing CNN architecture. To demonstrate the effectiveness of the proposed method on both the single-modality and cross-modality (Optical- InfraRed) near-duplicate image pair detection tasks, we conduct extensive experiments on three popular benchmark datasets, namely CaliforniaND (ND means near duplicate), Mir-Flickr Near Duplicate, and TNO Multi-band Image Data Collection. The experimental results show that the proposed method can achieve superior performance compared with many state-of-the-art methods on both tasks.

源语言英语
文章编号255
页(从-至)1-22
页数22
期刊Sensors
21
1
DOI
出版状态已出版 - 1 1月 2021

指纹

探究 'Single- and cross-modality near duplicate image pairs detection via spatial transformer comparing CNN' 的科研主题。它们共同构成独一无二的指纹。

引用此