Lip contour extraction method based on multiple active shape model for audio visual speech recognition

Lei Xie, Wei Feng, Rongchun Zhao

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

In audio visual speech recognition and lipreading, the widely used ASM (active shape model) for lip contour extraction suffers from the lack of robustness and cannot extract the exact lip contours due to the various mouth shape changes when uttering. We present a more robust model-multiple active shape model (MASM). The model classifies the mouth shapes into closed mouth set, half-opened mouth set, and round mouth set. An independent ASM is built for each different set with a tiny set of the training data. The MASM contour extraction algorithm automatically selects the best accurate lip contour from multiple shape searching procedures. Considering the consecutive changes of the mouth, a method for smoothing lip contours is also presented to correct the contour extraction errors. Experimental results from AVCONDIG database show that extraction accuracy achieved by the MASM is higher than that of conventional ASM 13%. The combination of the MASM and the contour-smoothing method leads to another 7% accuracy improvement. With the fusion of the exact lip contour feature and audio MFCC (mel frequency cepstral coefficients) feature, the average word recognition rates of the considered connected-digits speech recognition task are considerably increased under noisy acoustic conditions.

Original languageEnglish
Pages (from-to)674-678
Number of pages5
JournalXibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University
Volume22
Issue number5
StatePublished - Oct 2004

Keywords

  • Active shape model
  • Audio visual speech recognition
  • Lip contour extraction
  • Multiple active shape model
  • Speech recognition

Fingerprint

Dive into the research topics of 'Lip contour extraction method based on multiple active shape model for audio visual speech recognition'. Together they form a unique fingerprint.

Cite this