TY - GEN
T1 - Deception Detection Algorithm Based on Global and Local Feature Fusion with Multi-head Attention
AU - Kang, Jian
AU - Qu, Wen
AU - Cui, Shaoxing
AU - Feng, Xiaoyi
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Deception is a prevalent human behavior that significantly impacts our perception of essential facts. Therefore, developing accurate deception detection technology holds great significance. However, current research on pure visual deception detection algorithms does not leverage deep learning methods to extract detailed features such as facial Action Units (AUs) and Gaze Angles. Additionally, the global information within facial video sequences is often overlooked. To address these limitations, this paper introduces a novel deception detection model that combines global and local facial features through attention mechanisms. Firstly, the model focuses on the local features of the face, computing AU Strength and Gaze Angle for each frame to create a multivariate time series for every video. Subsequently, the Siamese Transformer model, employing Patching, extracts deep temporal and channel features from the multivariate time series. Additionally, the occurrence frequency of five specific AUs is selected as a manual feature. Secondly, the model conducts video understanding based on the global features of the face. Local features are extracted from each frame using Shallow CNNs with multiple sensitivity fields. Then, a Video Transformer model with spatiotemporal separation attention is applied to globally model the sequence of face frames. Finally, the extracted local and global facial features are concatenated and fed into a classifier to determine deception. Extensive experiments on existing datasets validate the outstanding performance of the proposed method.
AB - Deception is a prevalent human behavior that significantly impacts our perception of essential facts. Therefore, developing accurate deception detection technology holds great significance. However, current research on pure visual deception detection algorithms does not leverage deep learning methods to extract detailed features such as facial Action Units (AUs) and Gaze Angles. Additionally, the global information within facial video sequences is often overlooked. To address these limitations, this paper introduces a novel deception detection model that combines global and local facial features through attention mechanisms. Firstly, the model focuses on the local features of the face, computing AU Strength and Gaze Angle for each frame to create a multivariate time series for every video. Subsequently, the Siamese Transformer model, employing Patching, extracts deep temporal and channel features from the multivariate time series. Additionally, the occurrence frequency of five specific AUs is selected as a manual feature. Secondly, the model conducts video understanding based on the global features of the face. Local features are extracted from each frame using Shallow CNNs with multiple sensitivity fields. Then, a Video Transformer model with spatiotemporal separation attention is applied to globally model the sequence of face frames. Finally, the extracted local and global facial features are concatenated and fed into a classifier to determine deception. Extensive experiments on existing datasets validate the outstanding performance of the proposed method.
KW - Deception detection
KW - Facial AU
KW - Multivariate time series
KW - Transformer
KW - Video Understanding
UR - http://www.scopus.com/inward/record.url?scp=85199474577&partnerID=8YFLogxK
U2 - 10.1109/ICIPMC62364.2024.10586666
DO - 10.1109/ICIPMC62364.2024.10586666
M3 - 会议稿件
AN - SCOPUS:85199474577
T3 - 2024 3rd International Conference on Image Processing and Media Computing, ICIPMC 2024
SP - 162
EP - 168
BT - 2024 3rd International Conference on Image Processing and Media Computing, ICIPMC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 3rd International Conference on Image Processing and Media Computing, ICIPMC 2024
Y2 - 17 May 2024 through 19 May 2024
ER -