A novel design for a gated recurrent network with attentional memories

Libin Sun, Gao Biao, Haobin Shi

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Recurrent neural networks (RNN) are used for pro-cessing the temporal dynamics of complex sequential information. Existing recurrent neural networks, such as Long Short-Term Memory (LSTM) and its variant, Gated Recurrent Unit (GRU), are used to control current and historical information. This study proposes a novel design for a gated recurrent neural network with an attention mechanism, named the Attentional Memory Unit (AMU), which gives the recurrent neural network an attention capability. This design uses a gated mechanism and a generalized attention mechanism so that it can be adaptively adjusted in the spatial and temporal sense to allow the application of the attention mechanism for different scenarios. The experimental results verify that the proposed AMU significantly boosts the power of RNNs.

Original languageEnglish
Title of host publicationProceedings - 2022 International Conference on Computer Engineering and Artificial Intelligence, ICCEAI 2022
EditorsPan Lin, Yong Yang
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages522-527
Number of pages6
ISBN (Electronic)9781665468039
DOIs
StatePublished - 2022
Event2022 International Conference on Computer Engineering and Artificial Intelligence, ICCEAI 2022 - Shijiazhuang, China
Duration: 22 Jul 202224 Jul 2022

Publication series

NameProceedings - 2022 International Conference on Computer Engineering and Artificial Intelligence, ICCEAI 2022

Conference

Conference2022 International Conference on Computer Engineering and Artificial Intelligence, ICCEAI 2022
Country/TerritoryChina
CityShijiazhuang
Period22/07/2224/07/22

Keywords

  • Attention Mecha-nism
  • Recurrent Neural Network

Fingerprint

Dive into the research topics of 'A novel design for a gated recurrent network with attentional memories'. Together they form a unique fingerprint.

Cite this