Video summarization via block sparse dictionary selection

Mingyang Ma, Shaohui Mei, Shuai Wan, Junhui Hou, Zhiyong Wang, David Dagan Feng

Research output: Contribution to journalArticlepeer-review

71 Scopus citations

Abstract

The explosive growth of video data has raised new challenges for many video processing tasks such as video browsing and retrieval, hence, effective and efficient video summarization (VS) is urgently demanded to automatically summarize a video into a succinct version. Recent years have witnessed the advancements of sparse representation based approaches for VS. However, video frames are analyzed individually for keyframe selection in existing methods, which could lead to redundancy among selected keyframes and poor robustness to outlier frames. Due to that adjacent frames are visually similar, candidate keyframes often occur in temporal blocks, in addition to sparse presence. Therefore, in this paper, the block-sparsity of candidate keyframes is taken into consideration, by which the VS problem is formulated as a block sparse dictionary selection model. Moreover, a simultaneous block version of Orthogonal Matching Pursuit (SBOMP) algorithm is designed for model optimization. Two keyframe selection strategies are also explored for each block. Experimental results on two benchmark datasets, namely VSumm and TVSum datasets, demonstrate that the proposed SBOMP based VS method clearly outperforms several state-of-the-art sparse representation based methods in terms of F-score, redundancy among keyframes and robustness to outlier frames.

Original languageEnglish
Pages (from-to)197-209
Number of pages13
JournalNeurocomputing
Volume378
DOIs
StatePublished - 22 Feb 2020

Keywords

  • Block-sparsity
  • Dictionary selection
  • Sparse representation
  • Video summarization

Fingerprint

Dive into the research topics of 'Video summarization via block sparse dictionary selection'. Together they form a unique fingerprint.

Cite this