SAST: a suppressing ambiguity self-training framework for facial expression recognition

Zhe Guo, Bingxin Wei, Xuewen Liu, Zhibo Zhang, Shiya Liu, Yangyu Fan

Research output: Contribution to journalArticlepeer-review

Abstract

Facial expression recognition (FER) suffers from insufficient label information, as human expressions are complex and diverse, with many expressions ambiguous. Using low-quality labels or low-quantity labels will aggravate ambiguity of model predictions and reduce the accuracy of FER. How to improve the robustness of FER to ambiguous data with insufficient information remains challenging. To this end, we propose the Suppressing Ambiguity Self-Training (SAST) framework which is the first attempt to address the problem of insufficient information both label quality and label quantity containing, simultaneously. Specifically, we design an Ambiguous Relative Label Usage (ARLU) strategy that mixes hard labels and soft labels to alleviate the information loss problem caused by hard labels. We also enhance the robustness of the model to ambiguous data by means of Self-Training Resampling (STR). We further use the landmarks and Patch Branch (PB) to enhance the ability of suppressing ambiguity. Experiments on RAF-DB, FERPlus, SFEW, and AffectNet datasets show that our SAST outperforms 6 semi-supervised methods with fewer annotations, and achieves competitive accuracy to State-Of-The-Art (SOTA) FER methods. Our code is available at https://github.com/Liuxww/SAST.

Original languageEnglish
Pages (from-to)56059-56076
Number of pages18
JournalMultimedia Tools and Applications
Volume83
Issue number18
DOIs
StatePublished - May 2024

Keywords

  • Facial expression recognition
  • Insufficient information
  • Self-training
  • Suppressing ambiguity

Fingerprint

Dive into the research topics of 'SAST: a suppressing ambiguity self-training framework for facial expression recognition'. Together they form a unique fingerprint.

Cite this