TY - GEN
T1 - FMRI-Guided Time-Symmetric Joint Model for Visual Attention Prediction
AU - Wei, Yaonai
AU - Ma, Chong
AU - Zhong, Tianyang
AU - Du, Lei
AU - Zhang, Tuo
AU - Zhang, Songyao
AU - Yang, Li
AU - Liu, Tianming
AU - Zhang, Han
AU - He, Zhibin
AU - Shang, Muheng
AU - Han, Junwei
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Visual attention prediction is linked to brain activity, cognition, and behavior. Despite the availability of brain activity features, previous studies have not fully utilized them, resulting in saliency maps predicted by models primarily based on image features that do not accurately reflect visual attention in the human brain. This inspires us to use functional Magnetic Resonance Imaging (fMRI) signals as a 'brain observer' to supervise the training of developing models that integrate top-down image attention-dependent cues and supervise information from saliency maps generated from gaze movement patterns under natural stimuli. Hence, this paper presents an FMRI-Guided Time-Symmetric Joint Model to predict saliency maps from movie clips, which captures the dynamic aspects of human brain cognition and attention, enabling the combination of image features with brain features. Furthermore, we generalize the model to the MS-COCO challenge, evaluating its performance on non-movie data. Our model outperforms other brain-feature-free methods in focusing on visual attention regions of humans in both movie and non-movie datasets. Additionally, incorporating brain features improves model performance, indicating their ability to bridge the semantic gap between human cognition and visual images, allowing for more accurate capture of visual attention regions.
AB - Visual attention prediction is linked to brain activity, cognition, and behavior. Despite the availability of brain activity features, previous studies have not fully utilized them, resulting in saliency maps predicted by models primarily based on image features that do not accurately reflect visual attention in the human brain. This inspires us to use functional Magnetic Resonance Imaging (fMRI) signals as a 'brain observer' to supervise the training of developing models that integrate top-down image attention-dependent cues and supervise information from saliency maps generated from gaze movement patterns under natural stimuli. Hence, this paper presents an FMRI-Guided Time-Symmetric Joint Model to predict saliency maps from movie clips, which captures the dynamic aspects of human brain cognition and attention, enabling the combination of image features with brain features. Furthermore, we generalize the model to the MS-COCO challenge, evaluating its performance on non-movie data. Our model outperforms other brain-feature-free methods in focusing on visual attention regions of humans in both movie and non-movie datasets. Additionally, incorporating brain features improves model performance, indicating their ability to bridge the semantic gap between human cognition and visual images, allowing for more accurate capture of visual attention regions.
KW - fMRI
KW - gaze map
KW - saliency detection
UR - http://www.scopus.com/inward/record.url?scp=85184864690&partnerID=8YFLogxK
U2 - 10.1109/BIBM58861.2023.10385869
DO - 10.1109/BIBM58861.2023.10385869
M3 - 会议稿件
AN - SCOPUS:85184864690
T3 - Proceedings - 2023 2023 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2023
SP - 3212
EP - 3217
BT - Proceedings - 2023 2023 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2023
A2 - Jiang, Xingpeng
A2 - Wang, Haiying
A2 - Alhajj, Reda
A2 - Hu, Xiaohua
A2 - Engel, Felix
A2 - Mahmud, Mufti
A2 - Pisanti, Nadia
A2 - Cui, Xuefeng
A2 - Song, Hong
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2023 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2023
Y2 - 5 December 2023 through 8 December 2023
ER -