跳到主要导航 跳到搜索 跳到主要内容

MMGER: Multi-Modal and Multi-Granularity Generative Error Correction with LLM for Joint Accent and Speech Recognition

  • Bingshen Mu
  • , Xucheng Wan
  • , Naijun Zheng
  • , Huan Zhou
  • , Lei Xie
  • Northwestern Polytechnical University Xian
  • Huawei Technologies Co., Ltd.

科研成果: 期刊稿件文章同行评审

7 引用 (Scopus)

摘要

Despite notable advancements in automatic speech recognition (ASR), performance tends to degrade when faced with adverse conditions. Generative error correction (GER) leverages the exceptional text comprehension capabilities of large language models (LLM), delivering impressive performance in ASR error correction, where N-best hypotheses provide valuable information for transcription prediction. However, GER encounters challenges such as fixed N-best hypotheses, insufficient utilization of acoustic information, and limited specificity to multi-Accent scenarios. In this paper, we explore the application of GER in multi-Accent scenarios. Accents represent deviations from standard pronunciation norms, and the multi-Task learning framework for simultaneous ASR and accent recognition (AR) has effectively addressed the multi-Accent scenarios, making it a prominent solution. In this work, we propose a unified ASR-AR GER model, named MMGER, leveraging multi-modal correction, and multi-granularity correction. Multi-Task ASR-AR learning is employed to provide dynamic 1-best hypotheses and accent embeddings. Multi-modal correction accomplishes fine-grained frame-level correction by force-Aligning the acoustic features of speech with the corresponding character-level 1-best hypothesis sequence. Multi-granularity correction supplements the global linguistic information by incorporating regular 1-best hypotheses atop fine-grained multi-modal correction to achieve coarse-grained utterance-level correction. MMGER effectively mitigates the limitations of GER and tailors LLM-based ASR error correction for the multi-Accent scenarios. Experiments conducted on the multi-Accent Mandarin KeSpeech dataset demonstrate the efficacy of MMGER, achieving a 26.72% relative improvement in AR accuracy and a 27.55% relative reduction in ASR character error rate, compared to a well-established standard baseline.

源语言英语
页(从-至)1940-1944
页数5
期刊IEEE Signal Processing Letters
31
DOI
出版状态已出版 - 2024

指纹

探究 'MMGER: Multi-Modal and Multi-Granularity Generative Error Correction with LLM for Joint Accent and Speech Recognition' 的科研主题。它们共同构成独一无二的指纹。

引用此