Attention-guided dual spatial-temporal non-local network for video super-resolution

Wei Sun, Yanning Zhang

科研成果: 期刊稿件文章同行评审

14 引用 (Scopus)

摘要

In this paper, we propose an attention-guided dual spatial-temporal non-local network for video super-resolution (ADNLVSR). We integrate temporal and spatial non-local self-similar contexts from continuous video frames after motion compensation, and merge the features of different levels discriminatively with channel attention mechanism for target frame. During motion compensation, unlike previous methods directly stacking input images or features for merging, we use learnable attention mechanism to guide the merging, which suppresses undesired components caused by misalignment and enhances desirable fine details. During feature fusion, in contrast to most previous approaches where global-level non-local self-similarity existing in space or time is usually considered, we propose region-level spatial and temporal non-local operations for exploiting temporal correlations and enhancing similar spatial structures. The proposed modules can effectively avoid the computational burden caused by existing global-level non-local operations based on our analysis, and enhance correlated structure information. In addition, we propose a channel attention-guided residual dense block (CRDB), in which a second-order channel attention mechanism is applied to adaptively rescale the channel-wise features for more discriminative representations. Extensive experiments on different datasets demonstrate superior performance to state-of-the-art published methods on video super-resolution.

源语言英语
页(从-至)24-33
页数10
期刊Neurocomputing
406
DOI
出版状态已出版 - 17 9月 2020

指纹

探究 'Attention-guided dual spatial-temporal non-local network for video super-resolution' 的科研主题。它们共同构成独一无二的指纹。

引用此