TY - JOUR
T1 - Trademark image retrieval via transformation-invariant deep hashing
AU - Xia, Zhaoqiang
AU - Lin, Jie
AU - Feng, Xiaoyi
N1 - Publisher Copyright:
© 2019 Elsevier Inc.
PY - 2019/2
Y1 - 2019/2
N2 - Trademark images are usually used to distinguish goods due to their uniqueness, and the amount becomes too huge to search these images accurately and fast. Most existing methods utilize conventional dense features to search visually-similar images, however, the performance and search time are not satisfactory. In this paper, we propose a unified deep hashing framework to learn the binary codes for trademark images, resulting in good performance with less search time. The unified framework integrates two types of deep convolutional networks (i.e., spatial transformer network and recurrent convolutional network) for obtaining transformation-invariant features. These features are discriminative for describing trademark images and robust to different types of transformations. The two-stream networks are followed by the hashing layer. Network parameters are learned by minimizing a sample-weighted loss, which can leverage the hard-searched images. We conduct experiments on two benchmark image sets, i.e., NPU-TM and METU, and verify the effectiveness and efficiency of our proposed approach over state-of-the-art.
AB - Trademark images are usually used to distinguish goods due to their uniqueness, and the amount becomes too huge to search these images accurately and fast. Most existing methods utilize conventional dense features to search visually-similar images, however, the performance and search time are not satisfactory. In this paper, we propose a unified deep hashing framework to learn the binary codes for trademark images, resulting in good performance with less search time. The unified framework integrates two types of deep convolutional networks (i.e., spatial transformer network and recurrent convolutional network) for obtaining transformation-invariant features. These features are discriminative for describing trademark images and robust to different types of transformations. The two-stream networks are followed by the hashing layer. Network parameters are learned by minimizing a sample-weighted loss, which can leverage the hard-searched images. We conduct experiments on two benchmark image sets, i.e., NPU-TM and METU, and verify the effectiveness and efficiency of our proposed approach over state-of-the-art.
KW - Deep hashing
KW - Recurrent convolutional network
KW - Sample-weighted loss
KW - Spatial transformer network
KW - Trademark image retrieval
KW - Transformation-invariant feature
UR - http://www.scopus.com/inward/record.url?scp=85059760259&partnerID=8YFLogxK
U2 - 10.1016/j.jvcir.2019.01.011
DO - 10.1016/j.jvcir.2019.01.011
M3 - 文章
AN - SCOPUS:85059760259
SN - 1047-3203
VL - 59
SP - 108
EP - 116
JO - Journal of Visual Communication and Image Representation
JF - Journal of Visual Communication and Image Representation
ER -