跳到主要导航 跳到搜索 跳到主要内容

Revealing Event Saliency in Unconstrained Video Collection

  • Carnegie Mellon University
  • Northwestern Polytechnical University Xian

科研成果: 期刊稿件文章同行评审

123 引用 (Scopus)

摘要

Recent progresses in multimedia event detection have enabled us to find videos about a predefined event from a large-scale video collection. Research towards more intrinsic unsupervised video understanding is an interesting but understudied field. Specifically, given a collection of videos sharing a common event of interest, the goal is to discover the salient fragments, i.e., the curt video fragments that can concisely portray the underlying event of interest, from each video. To explore this novel direction, this paper proposes an unsupervised event saliency revealing framework. It first extracts features from multiple modalities to represent each shot in the given video collection. Then, these shots are clustered to build the cluster-level event saliency revealing framework, which explores useful information cues (i.e., the intra-cluster prior, inter-cluster discriminability, and inter-cluster smoothness) by a concise optimization model. Compared with the existing methods, our approach could highlight the intrinsic stimulus of the unseen event within a video in an unsupervised fashion. Thus, it could potentially benefit to a wide range of multimedia tasks like video browsing, understanding, and search. To quantitatively verify the proposed method, we systematically compare the method to a number of baseline methods on the TRECVID benchmarks. Experimental results have demonstrated its effectiveness and efficiency.

源语言英语
文章编号7835130
页(从-至)1746-1758
页数13
期刊IEEE Transactions on Image Processing
26
4
DOI
出版状态已出版 - 4月 2017

指纹

探究 'Revealing Event Saliency in Unconstrained Video Collection' 的科研主题。它们共同构成独一无二的指纹。

引用此