Title
Real-time 3D talking head from a synthetic viseme dataset
Abstract
In this paper, we describe a simple and fast way to build a 3D talking head which can be used in many applications requiring audiovisual speech animation system. The talking head is constructed from a synthetic 3D viseme dataset, which is realistic enough and can be generated with 3D modeling software. To build the talking head, at first the viseme dataset is analyzed statistically to obtain the optimal linear parameters to control the movements of the lips and jaw of the 3D head model. These parameters correspond to some of the low-level MPEG-4 FAPs, hence our method can be used to extract the speech-relevant MPEG-4 FAPs from a dataset of phonemes/visemes. The parameterized head model is eventually combined with a Text-to-Speech (TTS) system to synthesize audiovisual speech from a given text. To make the talking head looks more realistic, eye-blink and movements are also animated during the speech. We implemented this work in an interactive text-to-audio-visual speech system.
Year
DOI
Venue
2009
10.1145/1670252.1670260
VRCAI
Keywords
Field
DocType
audiovisual speech,synthetic viseme dataset,parameterized head model,audiovisual speech animation system,interactive text-to-audio-visual speech system,mpeg-4 faps,head model,optimal linear parameter,viseme dataset,text to speech,speech synthesis,real time,principal component analysis
Computer vision,Parameterized complexity,Computer science,Viseme,Animation system,Speech recognition,Software,Artificial intelligence,3D modeling
Conference
Citations 
PageRank 
References 
1
0.35
8
Authors
4
Name
Order
Citations
PageRank
Arthur Niswar1283.19
Ee Ping Ong231333.36
Hong Thai Nguyen351.95
Zhiyong Huang410611.79