Training augmentation with adversarial examples for robust speech recognition

Sining Sun, Ching Feng Yeh, Mari Ostendorf, Mei Yuh Hwang, Lei Xie

Research output: Contribution to journalConference articlepeer-review

51 Scopus citations

Abstract

This paper explores the use of adversarial examples in training speech recognition systems to increase robustness of deep neural network acoustic models. During training, the fast gradient sign method is used to generate adversarial examples augmenting the original training data. Different from conventional data augmentation based on data transformations, the examples are dynamically generated based on current acoustic model parameters. We assess the impact of adversarial data augmentation in experiments on the Aurora-4 and CHiME-4 single-channel tasks, showing improved robustness against noise and channel variation. Further improvement is obtained when combining adversarial examples with teacher/student training, leading to a 23% relative word error rate reduction on Aurora-4.

Original languageEnglish
Pages (from-to)2404-2408
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume2018-September
DOIs
StatePublished - 2018
Event19th Annual Conference of the International Speech Communication, INTERSPEECH 2018 - Hyderabad, India
Duration: 2 Sep 20186 Sep 2018

Keywords

  • Adversarial examples
  • Data augmentation
  • FGSM
  • Robust speech recognition
  • Teacher-student model

Fingerprint

Dive into the research topics of 'Training augmentation with adversarial examples for robust speech recognition'. Together they form a unique fingerprint.

Cite this