Learning Cross-Attention Discriminators via Alternating Time-Space Transformers for Visual Tracking

Wuwei Wang, Ke Zhang, Yu Su, Jingyu Wang, Qi Wang

科研成果: 期刊稿件文章同行评审

11 引用 (Scopus)

摘要

In the past few years, visual tracking methods with convolution neural networks (CNNs) have gained great popularity and success. However, the convolution operation of CNNs struggles to relate spatially distant information, which limits the discriminative power of trackers. Very recently, several Transformer-assisted tracking approaches have emerged to alleviate the above issue by combining CNNs with Transformers to enhance the feature representation. In contrast to the methods mentioned above, this article explores a pure Transformer-based model with a novel semi-Siamese architecture. Both the time-space self-attention module used to construct the feature extraction backbone and the cross-attention discriminator used to estimate the response map solely leverage attention without convolution. Inspired by the recent vision transformers (ViTs), we propose the multistage alternating time-space Transformers (ATSTs) to learn robust feature representation. Specifically, temporal and spatial tokens at each stage are alternately extracted and encoded by separate Transformers. Subsequently, a cross-attention discriminator is proposed to directly generate response maps of the search region without additional prediction heads or correlation filters. Experimental results show that our ATST-based model attains favorable results against state-of-the-art convolutional trackers. Moreover, it shows comparable performance with recent 'CNN + Transformer' trackers on various benchmarks while our ATST requires significantly less training data.

源语言英语
页(从-至)15156-15169
页数14
期刊IEEE Transactions on Neural Networks and Learning Systems
35
11
DOI
出版状态已出版 - 2024

指纹

探究 'Learning Cross-Attention Discriminators via Alternating Time-Space Transformers for Visual Tracking' 的科研主题。它们共同构成独一无二的指纹。

引用此