@inproceedings{3a2494f8f12841f089dd46d8da428aca,
title = "Robust MelGAN: A robust universal neural vocoder for high-fidelity TTS",
abstract = "In current two-stage neural text-to-speech (TTS) paradigm, it is ideal to have a universal neural vocoder, once trained, which is robust to imperfect mel-spectrogram predicted from the acoustic model. To this end, we propose Robust MelGAN vocoder by solving the original multi-band MelGAN's metallic sound problem and increasing its generalization ability. Specifically, we introduce a fine-grained network dropout strategy to the generator. With a specifically designed over-smooth handler which separates speech signal intro periodic and aperiodic components, we only perform network dropout to the aperodic components, which alleviates metallic sounding and maintains good speaker similarity. To further improve generalization ability, we introduce several data augmentation methods to augment fake data in the discriminator, including harmonic shift, harmonic noise and phase noise. Experiments show that Robust MelGAN can be used as a universal vocoder, significantly improving sound quality in TTS systems built on various types of data. 11Audio samples are available at https://RobustMelGAN.github.io/RobustMelGAN/",
keywords = "data augmentation, generative adversarial network, text-to speech, universal vocoder",
author = "Kun Song and Jian Cong and Xinsheng Wang and Yongmao Zhang and Lei Xie and Ning Jiang and Haiying Wu",
note = "Publisher Copyright: {\textcopyright} 2022 IEEE.; 13th International Symposium on Chinese Spoken Language Processing, ISCSLP 2022 ; Conference date: 11-12-2022 Through 14-12-2022",
year = "2022",
doi = "10.1109/ISCSLP57327.2022.10038120",
language = "英语",
series = "2022 13th International Symposium on Chinese Spoken Language Processing, ISCSLP 2022",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "71--75",
editor = "Lee, {Kong Aik} and Hung-yi Lee and Yanfeng Lu and Minghui Dong",
booktitle = "2022 13th International Symposium on Chinese Spoken Language Processing, ISCSLP 2022",
}