Summary on the Multimodal Information Based Speech Processing (MISP) 2022 Challenge

Hang Chen, Shilong Wu, Yusheng Dai, Zhe Wang, Jun Du, Chin Hui Lee, Jingdong Chen, Shinji Watanabe, Sabato Marco Siniscalchi, Odette Scharenborg, Di Yuan Liu, Bao Cai Yin, Jia Pan, Jian Qing Gao, Cong Liu

Research output: Contribution to journalConference articlepeer-review

1 Scopus citations

Abstract

The Multimodal Information based Speech Processing (MISP) 2022 challenge aimed to enhance speech processing performance in harsh acoustic environments by leveraging additional modalities such as video or text. The challenge included two tracks: audio-visual speaker diarization (AVSD) and audio-visual diarization and recognition (AVDR). The training material was based on previous MISP 2021 recordings, but we have accurately synchronized audio and visual data. Additionally, a new evaluation set was provided. This paper gives an overview of the challenge setup, presents the results, and summarizes the effective techniques employed by the participants. We also analyze the current technical challenges and suggest directions for future research in AVSD and AVDR.

Keywords

  • audio-visual
  • MISP challenge
  • speaker diarization
  • speech enhancement
  • speech recognition

Fingerprint

Dive into the research topics of 'Summary on the Multimodal Information Based Speech Processing (MISP) 2022 Challenge'. Together they form a unique fingerprint.

Cite this