HMM-based Automatic Visual Speech Segmentation Using Facial Data - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2010

HMM-based Automatic Visual Speech Segmentation Using Facial Data

Résumé

We describe automatic visual speech segmentation using facial data captured by a stereo-vision technique. The segmentation is performed using an HMM-based forced alignment mechanism widely used in automatic speech recognition. The idea is based on the assumption that using visual speech data alone for the training might capture the uniqueness in the facial compo- nent of speech articulation, asynchrony (time lags) in visual and acoustic speech segments and significant coarticulation effects. This should provide valuable information that helps to show the extent to which a phoneme may affect surrounding phonemes visually. This should provide information valuable in labeling the visual speech segments based on dominant coarticulatory contexts.
Fichier principal
Vignette du fichier
IS10-UM.pdf (217.75 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

inria-00526776 , version 1 (15-10-2010)

Identifiants

  • HAL Id : inria-00526776 , version 1

Citer

Utpala Musti, Asterios Toutios, Slim Ouni, Vincent Colotte, Brigitte Wrobel-Dautcourt, et al.. HMM-based Automatic Visual Speech Segmentation Using Facial Data. Interspeech 2010, ISCA, Sep 2010, Makuhari, Chiba, Japan. pp.1401-1404. ⟨inria-00526776⟩
362 Consultations
253 Téléchargements

Partager

Gmail Facebook X LinkedIn More