Speech animation using coupled hidden Markov models

Lei Xie, Zhi Qiang Liu

科研成果: 书/报告/会议事项章节会议稿件同行评审

14 引用 (Scopus)

摘要

We present a novel speech animation approach using coupled hidden Markov models (CHMMs). Different from the conventional HMMs that use a single state chain to model the audio-visual speech with tight inter-modal synchronization, we use the CHMMs to model the asynchrony, different discriminative abilities, and temporal coupling between the audio speech and the visual speech, which are important factors for animations looking natural. Based on the audio-visual CHMMs, visual animation parameters are predicted from audio through an EM-based audio to visual conversion algorithm. Experiments on the JEWEL AV database show that compared with the conventional HMMs, the CHMMs can output visual parameters that are much closer to the actual ones. Explicit modelling of audio-visual speech is promising in speech animation.

源语言英语
主期刊名Proceedings - 18th International Conference on Pattern Recognition, ICPR 2006
1128-1131
页数4
DOI
出版状态已出版 - 2006
已对外发布
活动18th International Conference on Pattern Recognition, ICPR 2006 - Hong Kong, 中国
期限: 20 8月 200624 8月 2006

出版系列

姓名Proceedings - International Conference on Pattern Recognition
1
ISSN(印刷版)1051-4651

会议

会议18th International Conference on Pattern Recognition, ICPR 2006
国家/地区中国
Hong Kong
时期20/08/0624/08/06

引用此