跳到主要导航 跳到搜索 跳到主要内容

Spatial and temporal visual attention prediction in videos using eye movement data

  • Junwei Han
  • , Liye Sun
  • , Xintao Hu
  • , Jungong Han
  • , Ling Shao
  • Northwestern Polytechnical University Xian
  • Civolution Technology
  • University of Sheffield

科研成果: 期刊稿件文章同行评审

27 引用 (Scopus)

摘要

Visual attention detection in static images has achieved outstanding progress in recent years whereas much less effort has been devoted to learning visual attention in video sequences. In this paper, we propose a novel method to model spatial and temporal visual attention for videos respectively through learning from human gaze data. The spatial visual attention mainly predicts where viewers look in each video frame while the temporal visual attention measures which video frame is more likely to attract viewers' interest. Our underlying premise is that objects as well as their movements, instead of conventional contrast-related information, are major factors in dynamic scenes to drive visual attention. Firstly, the proposed models extract two types of bottom-up features derived from multi-scale object filter responses and spatiotemporal motion energy, respectively. Then, spatiotemporal gaze density and inter-observer gaze congruency are generated using a large collection of human-eye gaze data to form two training sets. Finally, prediction models of temporal visual attention and spatial visual attention are learned based on those two training sets and bottom-up features, respectively. Extensive evaluations on publicly available video benchmarks and applications in interestingness prediction of movie trailers demonstrate the effectiveness of the proposed work.

源语言英语
页(从-至)140-153
页数14
期刊Neurocomputing
145
DOI
出版状态已出版 - 5 12月 2014

指纹

探究 'Spatial and temporal visual attention prediction in videos using eye movement data' 的科研主题。它们共同构成独一无二的指纹。

引用此