An adaptive multi-sensor visual attention model

Wenbai Chen, Jingchen Li, Haobin Shi, Kao Shing Hwang

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

The emerging recurrent visual attention models mostly utilize a sensor to continuously capture features from the input, which requires a suited design for the sensor. Researchers usually need a number of attempts to determine optimal structures for the sensor and corresponding modules. In this work, an adaptive multi-sensor visual attention model (AM-MA) is proposed to enhance the recurrent visual attention model. The proposed model uses several sensors to observe the original input recurrently, while the number of sensors can be added adaptively. Each sensor generates a hidden state and is followed by a location network to provide the deployment scheme. We design a self-evaluation mechanism for AM-MA, by which it can decide whether to add new sensors during training. Besides, the proposed AM-MA leverages a fine-tune mechanism to avoid a lengthy training process. AM-MA is a parameter-insensitive model. That is, there is no need for researchers to pre-train the model for finding the optimal structure in the case of unknown complexity. Experimental results show that the proposed AM-MA not only outperforms the renowned sensor-based attention model on image classification tasks, but also achieves satisfactory results when given an inappropriate structure.

Original languageEnglish
Pages (from-to)7241-7252
Number of pages12
JournalNeural Computing and Applications
Volume34
Issue number9
DOIs
StatePublished - May 2022

Keywords

  • Attention mechanism
  • Neural network
  • Visual attention model

Fingerprint

Dive into the research topics of 'An adaptive multi-sensor visual attention model'. Together they form a unique fingerprint.

Cite this