SELM: SPEECH ENHANCEMENT USING DISCRETE TOKENS AND LANGUAGE MODELS

Ziqian Wang, Xinfa Zhu, Zihan Zhang, Yuan Jun Lv, Ning Jiang, Guoqing Zhao, Lei Xie

Research output: Contribution to journalConference articlepeer-review

5 Scopus citations

Abstract

Language models (LMs) have recently shown superior performances in various speech generation tasks, demonstrating their powerful ability for semantic context modeling. Given the intrinsic similarity between speech generation and speech enhancement, harnessing semantic information is advantageous for speech enhancement tasks. In light of this, we propose SELM, a novel speech enhancement paradigm that integrates discrete tokens and leverages language models. SELM comprises three stages: encoding, modeling, and decoding. We transform continuous waveform signals into discrete tokens using pre-trained self-supervised learning (SSL) models and a k-means tokenizer. Language models then capture comprehensive contextual information within these tokens. Finally, a de-tokenizer and HiFi-GAN restore them into enhanced speech. Experimental results demonstrate that SELM achieves comparable performance in objective metrics and superior subjective perception results. Our demos are available.

Original languageEnglish
Pages (from-to)11561-11565
Number of pages5
JournalICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
DOIs
StatePublished - 2024
Event2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Seoul, Korea, Republic of
Duration: 14 Apr 202419 Apr 2024

Keywords

  • generative model
  • language models
  • speech enhancement
  • staged approach

Fingerprint

Dive into the research topics of 'SELM: SPEECH ENHANCEMENT USING DISCRETE TOKENS AND LANGUAGE MODELS'. Together they form a unique fingerprint.

Cite this