TY - GEN
T1 - Bridging low-level features and high-level semantics via fMRI brain imaging for video classification
AU - Hu, Xintao
AU - Deng, Fan
AU - Li, Kaiming
AU - Zhang, Tuo
AU - Chen, Hanbo
AU - Jiang, Xi
AU - Lv, Jinglei
AU - Zhu, Dajiang
AU - Faraco, Carlos
AU - Zhang, Degang
AU - Mesbah, Arsham
AU - Han, Junwei
AU - Hua, Xiansheng
AU - Xie, Li
AU - Miller, Stephen
AU - Guo, Lei
AU - Liu, Tianming
PY - 2010
Y1 - 2010
N2 - The multimedia content analysis community has made significant effort to bridge the gap between low-level features and high-level semantics perceived by human cognitive systems such as real-world objects and concepts. In the two fields of multimedia analysis and brain imaging, both topics of low-level features and high level semantics are extensively studied. For instance, in the multimedia analysis field, many algorithms are available for multimedia feature extraction, and benchmark datasets are available such as the TRECVID. In the brain imaging field, brain regions that are responsible for vision, auditory perception, language, and working memory are well studied via functional magnetic resonance imaging (fMRI). This paper presents our initial effort in marrying these two fields in order to bridge the gaps between low-level features and high-level semantics via fMRI brain imaging. Our experimental paradigm is that we performed fMRI brain imaging when university student subjects watched the video clips selected from the TRECVID datasets. At current stage, we focus on the three concepts of sports, weather, and commercial-/advertisement specified in the TRECVID 2005. Meanwhile, the brain regions in vision, auditory, language, and working memory networks are quantitatively localized and mapped via task-based paradigm fMRI, and the fMRI responses in these regions are used to extract features as the representation of the brain's comprehension of semantics. Our computational framework aims to learn the most relevant low-level feature sets that best correlate the fMRI-derived semantics based on the training videos with fMRI scans, and then the learned models are applied to larger scale test datasets without fMRI scans for category classifications. Our result shows that: 1) there are meaningful couplings between brain's fMRI responses and video stimuli, suggesting the validity of linking semantics and low-level features via fMRI; 2) The computationally learned low-level feature sets from fMRI-derived semantic features can significantly improve the classification of video categories in comparison with that based on original low-level features.
AB - The multimedia content analysis community has made significant effort to bridge the gap between low-level features and high-level semantics perceived by human cognitive systems such as real-world objects and concepts. In the two fields of multimedia analysis and brain imaging, both topics of low-level features and high level semantics are extensively studied. For instance, in the multimedia analysis field, many algorithms are available for multimedia feature extraction, and benchmark datasets are available such as the TRECVID. In the brain imaging field, brain regions that are responsible for vision, auditory perception, language, and working memory are well studied via functional magnetic resonance imaging (fMRI). This paper presents our initial effort in marrying these two fields in order to bridge the gaps between low-level features and high-level semantics via fMRI brain imaging. Our experimental paradigm is that we performed fMRI brain imaging when university student subjects watched the video clips selected from the TRECVID datasets. At current stage, we focus on the three concepts of sports, weather, and commercial-/advertisement specified in the TRECVID 2005. Meanwhile, the brain regions in vision, auditory, language, and working memory networks are quantitatively localized and mapped via task-based paradigm fMRI, and the fMRI responses in these regions are used to extract features as the representation of the brain's comprehension of semantics. Our computational framework aims to learn the most relevant low-level feature sets that best correlate the fMRI-derived semantics based on the training videos with fMRI scans, and then the learned models are applied to larger scale test datasets without fMRI scans for category classifications. Our result shows that: 1) there are meaningful couplings between brain's fMRI responses and video stimuli, suggesting the validity of linking semantics and low-level features via fMRI; 2) The computationally learned low-level feature sets from fMRI-derived semantic features can significantly improve the classification of video categories in comparison with that based on original low-level features.
KW - brain computer interface
KW - brain imaging
KW - high-level features
KW - human vision
KW - low-level features
KW - semantics
UR - http://www.scopus.com/inward/record.url?scp=78650971374&partnerID=8YFLogxK
U2 - 10.1145/1873951.1874016
DO - 10.1145/1873951.1874016
M3 - 会议稿件
AN - SCOPUS:78650971374
SN - 9781605589336
T3 - MM'10 - Proceedings of the ACM Multimedia 2010 International Conference
SP - 451
EP - 460
BT - MM'10 - Proceedings of the ACM Multimedia 2010 International Conference
T2 - 18th ACM International Conference on Multimedia ACM Multimedia 2010, MM'10
Y2 - 25 October 2010 through 29 October 2010
ER -