TY - GEN
T1 - Improving Multimodal Emotion Recognition by Leveraging Acoustic Adaptation and Visual Alignment
AU - Zhao, Zhixian
AU - Chen, Haifeng
AU - Li, Xi
AU - Jiang, Dongmei
AU - Xie, Lei
N1 - Publisher Copyright:
© 2024 ACM.
PY - 2024/10/28
Y1 - 2024/10/28
N2 - Multimodal Emotion Recognition (MER) aims to automatically identify and understand human emotional states by integrating information from various modalities. However, the scarcity of annotated multimodal data significantly hinders the advancement of this research field. This paper presents our solution for the MER-SEMI sub-challenge of MER 2024. First, to better adapt acoustic modality features for the MER task, we experimentally evaluate the contributions of different layers of the pre-trained speech model HuBERT in emotion recognition. Based on these observations, we perform Parameter-Efficient Fine-Tuning (PEFT) on the layers identified as most effective for emotion recognition tasks, thereby achieving optimal adaptation for emotion recognition with a minimal number of learnable parameters. Second, leveraging the strengths of the acoustic modality, we propose a feature alignment pre-training method. This approach uses large-scale unlabeled data to train a visual encoder, thereby promoting the semantic alignment of visual features within the acoustic feature space. Finally, using the adapted acoustic features, aligned visual features, and lexical features, we employ an attention mechanism for feature fusion. On the MER2024-SEMI test set, the proposed method achieves a weighted F1 score of 88.90%, ranking fourth among all participating teams, validating the effectiveness of our approach.
AB - Multimodal Emotion Recognition (MER) aims to automatically identify and understand human emotional states by integrating information from various modalities. However, the scarcity of annotated multimodal data significantly hinders the advancement of this research field. This paper presents our solution for the MER-SEMI sub-challenge of MER 2024. First, to better adapt acoustic modality features for the MER task, we experimentally evaluate the contributions of different layers of the pre-trained speech model HuBERT in emotion recognition. Based on these observations, we perform Parameter-Efficient Fine-Tuning (PEFT) on the layers identified as most effective for emotion recognition tasks, thereby achieving optimal adaptation for emotion recognition with a minimal number of learnable parameters. Second, leveraging the strengths of the acoustic modality, we propose a feature alignment pre-training method. This approach uses large-scale unlabeled data to train a visual encoder, thereby promoting the semantic alignment of visual features within the acoustic feature space. Finally, using the adapted acoustic features, aligned visual features, and lexical features, we employ an attention mechanism for feature fusion. On the MER2024-SEMI test set, the proposed method achieves a weighted F1 score of 88.90%, ranking fourth among all participating teams, validating the effectiveness of our approach.
KW - contrastive learning
KW - fine-tuning
KW - multimodal emotion recognition
UR - http://www.scopus.com/inward/record.url?scp=85210856347&partnerID=8YFLogxK
U2 - 10.1145/3689092.3689407
DO - 10.1145/3689092.3689407
M3 - 会议稿件
AN - SCOPUS:85210856347
T3 - MRAC 2024 - Proceedings of the 2nd International Workshop on Multimodal and Responsible Affective Computing
SP - 67
EP - 71
BT - MRAC 2024 - Proceedings of the 2nd International Workshop on Multimodal and Responsible Affective Computing
PB - Association for Computing Machinery, Inc
T2 - 2nd International Workshop on Multimodal and Responsible Affective Computing, MRAC 2024
Y2 - 28 October 2024 through 1 November 2024
ER -