A Fast Algorithm for Lip Contour and Feature Extraction

Article Preview

Abstract:

The paper proposes a fast method that uses the lips color information in the Lab color space and the pre-knowledge of geometric characteristics around lip areas to extract the lip contour and visual speech features from color images or video sequences with front talking faces. In our method, the Adaboost algorithm is utilized to realize the face detection. Then, the mouth area is segmented based on the face shape attribution. According to the relative position of the trough and crest of the histogram, we can get an adaptive threshold. The A-component in the Lab color space was used to extract the outer lip and the L-component is used to extract the inner lip. From the contour image, we obtain the feature by searching twice the points of the contour. The experimental results show that obtained visual feature values in our approach are approximate to that with AAM algorithm but with less computation complexity.

You might also be interested in these eBooks

Info:

Periodical:

Pages:

2040-2044

Citation:

Online since:

June 2012

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2012 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

[1] G. Potamianos, C. Neti, G. Gravier, A. Garg, and A. Senior: Recent Advances in the Automatic Recognition of Audio-Visual Speech, In Proc. IEEE, 1306-1326 (2004).

DOI: 10.1109/jproc.2003.817150

Google Scholar

[2] J. -W. Kuo, H. -Y. Lo, and H. -M. Wang: Improved HMM/SVM methods for automatic phoneme segmentation, in Proc. Interspeech, Antwerp, Belgium, 2057-2060 (2007).

DOI: 10.21437/interspeech.2007-557

Google Scholar

[3] S. Gurbuz, Z. Tufekci, E. Patterson, and J. N. Gowdy: Application of affine-invariant Fourier descriptors to lipreading for audio-visual speech recognition, in Proc. Int. Conf. Acoustics, Speech, and Signal Processing, 177–180(2001).

DOI: 10.1109/icassp.2001.940796

Google Scholar

[4] S. Dupont and J. Luettin: Audio-visual speech modeling for continuous speech recognition, IEEE Trans. Multimedia, Vol. 2, p.141–151 (2000).

DOI: 10.1109/6046.865479

Google Scholar

[5] A. V. Nefian, L. Liang, X. Pi, X. Liu, and K. Murphy: Dynamic Bayesian networks for audio-visual speech recognition, EURASIP J. Appl. Signal Process., p.1274–1288 (2002).

DOI: 10.1155/s1110865702206083

Google Scholar

[6] Qing-Cai Chen, Guang-Hong Deng, Xiao-Long Wang, He-Jiao Huang: An Inner Contour Based Lip Moving Feature Extraction Method for Chinese Speech, Machine Learning and Cybernetics, International Conference, pp.3859-3864(2006).

DOI: 10.1109/icmlc.2006.258735

Google Scholar

[7] Ha-sung Koo, Ho-gen Song: Facial Feature Extraction for face Modeling Program, International Journal Of Circuits, Systems And Signal Processing, Issue 4, volume 4, pp.169-176(2010).

Google Scholar

[8] Takeshi Saitoh, Kazutoshi Morishita and Ryosuke Konishi: Analysis of Efficient Lip Reading Method for Various Languages, Pattern Recognition, 19th International Conference, pp.1-4(2008).

DOI: 10.1109/icpr.2008.4761049

Google Scholar