Context dependent viseme models for voice driven animation

Xie Lei, Jiang Dongmei, I. Ravyse, W. Verhelst, H. Sahli, V. Slavova, Z. Rongchun

科研成果: 书/报告/会议事项章节会议稿件同行评审

7 引用 (Scopus)

摘要

This paper addresses the problem of animating a talking figure, such as an avatar, using speech input only. The system that was developed is based on hidden Markov models for the acoustic observation vectors of the speech sounds that correspond to each of 16 visually distinct mouth shapes (visemes). The acoustic variability with context was taken into account by building acoustic viseme models that are dependent on the left and right viseme contexts. Our experimental results show that it is indeed possible to obtain visually relevant speech segmentation data directly from the purely acoustic speech signal.

源语言英语
主期刊名Proceedings EC-VIP-MC 2003 - 4th EURASIP Conference Focused on Video / Image Processing and Multimedia Communications
编辑Sonja Grgic, Mislav Grgic
出版商Institute of Electrical and Electronics Engineers Inc.
649-654
页数6
ISBN(电子版)9531840547, 9789531840545
DOI
出版状态已出版 - 2003
活动4th EURASIP Conference Focused on Video / Image Processing and Multimedia Communications, EC-VIP-MC 2003 - Zagreb, 克罗地亚
期限: 2 7月 20035 7月 2003

出版系列

姓名Proceedings EC-VIP-MC 2003 - 4th EURASIP Conference Focused on Video / Image Processing and Multimedia Communications
2

会议

会议4th EURASIP Conference Focused on Video / Image Processing and Multimedia Communications, EC-VIP-MC 2003
国家/地区克罗地亚
Zagreb
时期2/07/035/07/03

指纹

探究 'Context dependent viseme models for voice driven animation' 的科研主题。它们共同构成独一无二的指纹。

引用此