TY - JOUR
T1 - A Comparative Study on Speaker-attributed Automatic Speech Recognition in Multi-party Meetings
AU - Yu, Fan
AU - Du, Zhihao
AU - Zhang, Shiliang
AU - Lin, Yuxiao
AU - Xie, Lei
N1 - Publisher Copyright:
Copyright © 2022 ISCA.
PY - 2022
Y1 - 2022
N2 - In this paper, we conduct a comparative study on speaker-attributed automatic speech recognition (SA-ASR) in the multiparty meeting scenario, a topic with increasing attention in meeting rich transcription. Specifically, three approaches are evaluated in this study. The first approach, FD-SOT, consists of a frame-level diarization model to identify speakers and a multi-talker ASR to recognize utterances. The speaker-attributed transcriptions are obtained by aligning the diarization results and the recognized hypotheses. However, due to the modular independence, such an alignment strategy may suffer from erroneous timestamps which severely hinder the model performance. Therefore, we propose the second approach, WD-SOT, to address alignment errors by introducing a word-level diarization model, which can get rid of such timestamp alignment dependency. To further mitigate the alignment issues, we propose the third approach, TS-ASR, which trains a target-speaker separation module and an ASR module jointly. By comparing various strategies for each SA-ASR approach, experimental results on a real meeting scenario corpus, AliMeeting, reveal that the WD-SOT approach achieves 10.7% relative reduction on averaged speaker-dependent character error rate (SD-CER), compared with the FD-SOT approach. In addition, the TS-ASR approach also outperforms the FD-SOT approach and brings 16.5% relative average SD-CER reduction.
AB - In this paper, we conduct a comparative study on speaker-attributed automatic speech recognition (SA-ASR) in the multiparty meeting scenario, a topic with increasing attention in meeting rich transcription. Specifically, three approaches are evaluated in this study. The first approach, FD-SOT, consists of a frame-level diarization model to identify speakers and a multi-talker ASR to recognize utterances. The speaker-attributed transcriptions are obtained by aligning the diarization results and the recognized hypotheses. However, due to the modular independence, such an alignment strategy may suffer from erroneous timestamps which severely hinder the model performance. Therefore, we propose the second approach, WD-SOT, to address alignment errors by introducing a word-level diarization model, which can get rid of such timestamp alignment dependency. To further mitigate the alignment issues, we propose the third approach, TS-ASR, which trains a target-speaker separation module and an ASR module jointly. By comparing various strategies for each SA-ASR approach, experimental results on a real meeting scenario corpus, AliMeeting, reveal that the WD-SOT approach achieves 10.7% relative reduction on averaged speaker-dependent character error rate (SD-CER), compared with the FD-SOT approach. In addition, the TS-ASR approach also outperforms the FD-SOT approach and brings 16.5% relative average SD-CER reduction.
KW - AliMeeting
KW - multi-speaker ASR
KW - rich transcription
KW - speaker-attributed
UR - http://www.scopus.com/inward/record.url?scp=85140048979&partnerID=8YFLogxK
U2 - 10.21437/Interspeech.2022-11210
DO - 10.21437/Interspeech.2022-11210
M3 - 会议文章
AN - SCOPUS:85140048979
SN - 2308-457X
VL - 2022-September
SP - 560
EP - 564
JO - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
JF - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
T2 - 23rd Annual Conference of the International Speech Communication Association, INTERSPEECH 2022
Y2 - 18 September 2022 through 22 September 2022
ER -