跳到主要导航 跳到搜索 跳到主要内容

Two-Stage Learning to Predict Human Eye Fixations via SDAEs

科研成果: 期刊稿件文章同行评审

124 引用 (Scopus)

摘要

Saliency detection models aiming to quantitatively predict human eye-Attended locations in the visual field have been receiving increasing research interest in recent years. Unlike traditional methods that rely on hand-designed features and contrast inference mechanisms, this paper proposes a novel framework to learn saliency detection models from raw image data using deep networks. The proposed framework mainly consists of two learning stages. At the first learning stage, we develop a stacked denoising autoencoder (SDAE) model to learn robust, representative features from raw image data under an unsupervised manner. The second learning stage aims to jointly learn optimal mechanisms to capture the intrinsic mutual patterns as the feature contrast and to integrate them for final saliency prediction. Given the input of pairs of a center patch and its surrounding patches represented by the features learned at the first stage, a SDAE network is trained under the supervision of eye fixation labels, which achieves both contrast inference and contrast integration simultaneously. Experiments on three publically available eye tracking benchmarks and the comparisons with 16 state-of-The-Art approaches demonstrate the effectiveness of the proposed framework.

源语言英语
文章编号7051244
页(从-至)487-498
页数12
期刊IEEE Transactions on Cybernetics
46
2
DOI
出版状态已出版 - 2月 2016

指纹

探究 'Two-Stage Learning to Predict Human Eye Fixations via SDAEs' 的科研主题。它们共同构成独一无二的指纹。

引用此