A Cantonese speech-driven talking face using translingual audio-to-visual conversion

Lei Xie, Helen Meng, Zhi Qiang Liu

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

This paper proposes a novel approach towards a video- realistic, speech-driven talking face for Cantonese. We present a technique that realizes a talking face for a target language (Cantonese) using only audio-visual facial recordings for a base language (English). Given a Cantonese speech input, we first use a Cantonese speech recognizer to generate a Cantonese syllable transcription. Then we map it to an English phoneme transcription via a translingual mapping scheme that involves symbol mapping and time alignment from Cantonese syllables to English phonemes. With the phoneme transcription, the input speech, and the audio-visual models for English, an EM-based conversion algorithm is adopted to generate mouth animation parameters associated with the input Cantonese audio. We have carried out audio-visual syllable recognition experiments to objectively evaluate the proposed talking face. Results show that the visual speech synthesized by the Cantonese talking face can effectively increase the accuracy of Cantonese syllable recognition under noisy acoustic conditions.

源语言英语
主期刊名Chinese Spoken Language Processing - 5th International Symposium, ISCSLP 2006, Proceedings
627-639
页数13
DOI
出版状态已出版 - 2006
已对外发布
活动5th International Symposium on Chinese Spoken Language Processing, ISCSLP 2006 - Singapore, 新加坡
期限: 13 12月 200616 12月 2006

出版系列

姓名Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
4274 LNAI
ISSN(印刷版)0302-9743
ISSN(电子版)1611-3349

会议

会议5th International Symposium on Chinese Spoken Language Processing, ISCSLP 2006
国家/地区新加坡
Singapore
时期13/12/0616/12/06

指纹

探究 'A Cantonese speech-driven talking face using translingual audio-to-visual conversion' 的科研主题。它们共同构成独一无二的指纹。

引用此