An articulatory approach to video-realistic mouth animation

Lei Xie, Zhi Qiang Liu

科研成果: 书/报告/会议事项章节会议稿件同行评审

4 引用 (Scopus)

摘要

We propose an articulatory approach which is capable of converting speaker independent continuous speech into video-realistic mouth animation. We directly model the motions of articulators, such as lips, tongue, and teeth, using a Dynamic Bayesian Network (DBN)-structured articulatory model (AM). We also present an EM-based conversion algorithm to convert audio to animation parameters by maximizing the likelihood of these parameters given the input audio and the AMs. We further extend the AMs with introduction of speech context information, resulting in context dependent articulatory models (CD-AMs). Objective evaluations on the JEWEL testing set show that the animation parameters estimated by the proposed AMs and CD-AMs can follow the real parameters more accurately than that of phoneme-based models (PMs) and their context dependent counterparts (CD-PMs). Subjective evaluations on an AV subjective testing set, which collects various AV contents from the Internet, also demonstrate that the AMs and CD-AMs are able to generate more natural and realistic mouth animations and the CD-AMs achieve the best performance.

源语言英语
主期刊名2006 IEEE International Conference on Acoustics, Speech, and Signal Processing - Proceedings
I593-I596
出版状态已出版 - 2006
已对外发布
活动2006 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2006 - Toulouse, 法国
期限: 14 5月 200619 5月 2006

出版系列

姓名ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
1
ISSN(印刷版)1520-6149

会议

会议2006 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2006
国家/地区法国
Toulouse
时期14/05/0619/05/06

指纹

探究 'An articulatory approach to video-realistic mouth animation' 的科研主题。它们共同构成独一无二的指纹。

引用此