Study on Dual Mode Fusion Method of Video and Audio

Article Preview

Abstract:

In order to solve the hearing-impaired students in class only rely on sign language, amount of classroom information received less, This paper studies video and audio dual mode fusion algorithm combined with lip reading、speech recognition technology and information fusion technology.First ,speech feature extraction, processing of speech signal, the speech synchronization output text. At the same time, extraction of video features, voice and video signal fusion, Make voice information into visual information that the hearing-impaired students can receive. Make the students receive text messages as receive visual information, improve speech recognition rate, so meet the need of the classroom teaching for hearing-impaired students.

You might also be interested in these eBooks

Info:

Periodical:

Pages:

412-415

Citation:

Online since:

February 2015

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2015 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

* - Corresponding Author

[1] B. Mark. E.J. Holden, Robyn Owens: Lip tracking using pattern matching snakes : ACCV2002 Publications, Australia. 2002. pp.1-6.

Google Scholar

[2] Quoc Dinh Nguyen , Maurice Milgram, Thi-Hoang-Lan Nguyen: Multi Features Models for Robust Lip Tracking: 2008 10th Intl. Conf. on Control, Automation, Robotics and Vision, Hanoi, Vietnam, 2008. pp.1333-1337.

DOI: 10.1109/icarcv.2008.4795715

Google Scholar

[3] Brice Beaumesnil , Franck Luthon: Real Time Tracking for 3D Realistic Lip Animation, Proceedings of the 18th International Conference on Pattern Recognition, Norwich, NY (2006).

DOI: 10.1109/icpr.2006.956

Google Scholar

[4] Zhiyan Han: Research on Robust Feature Extracting and Visualization of Speech Signal, Doctor, Northeastern University, 2009. In Chinese.

Google Scholar

[5] Lirong Wang, Jianlei. Wang: Lip contour modeling based on PCA. Optics and Precision Engineering Vol. 20 (2012) , pp.2775-2779. In Chinese.

Google Scholar