Lip shape and hand position fusion for automatic vowel recognition in Cued Speech for French

Fiche du document

Date

2009

Type de document
Périmètre
Langue
Identifiants
Relations

Ce document est lié à :
info:eu-repo/semantics/altIdentifier/doi/10.1109/LSP.2009.2016011

Collection

Archives ouvertes

Licence

info:eu-repo/semantics/OpenAccess



Sujets proches En

Talking Melting

Citer ce document

Panikos Heracleous et al., « Lip shape and hand position fusion for automatic vowel recognition in Cued Speech for French », HAL-SHS : sciences de l'information, de la communication et des bibliothèques, ID : 10.1109/LSP.2009.2016011


Métriques


Partage / Export

Résumé En

Cued Speech is a sound-based system, which uses handshapes in different positions and in combination with lippatterns of speech, and makes all the sounds of spoken language clearly understandable to deaf and hearing-impaired people. The aim of Cued Speech is to overcome the problems of lip reading and thus enable deaf children and adults to wholly understand spoken language. Cued Speech recognition requires gesture recognition and lip shape recognition. In addition, the integration of the two components is of great importance. This article presents hidden Markov model (HMM)-based vowel recognition as used in Cued Speech for French. Based on concatenative feature fusion and multi-stream HMM decision fusion, lip shape and hand position components were integrated into a single component, and automatic vowel recognition was realized. In the case of multi-stream HMM decision fusion, the obtained vowel classication accuracy using lip shape and hand position information was 87.6%, showing absolute improvement of 19.6% in comparison with a use restricted only to lip parameters. The results achieved show the effectiveness of the proposed approaches to Cued Speech recognition.

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Sur les mêmes disciplines

Exporter en