Abstract | ||
---|---|---|
We aim at the realization of an Embodied Conversational Agent able to interact naturally and emotionally with user. In particular, the agent should behave expressively. Specifying for a given emotion, its corresponding facial expression will not produce the sensation of expressivity. To do so, one needs to specify parameters such as intensity, tension, movement property. Moreover, emotion affects also lip shapes during speech. Simply adding the facial expression of emotion to the lip shape does not produce lip readable movement. In this paper we present a model based on real data from a speaker on which was applied passive markers. The real data covers natural speech as well as emotional speech. We present an algorithm that determines the appropriate viseme and applies coarticulation and correlation rules to consider the vocalic and the consonantal contexts as well as muscular phenomena such as lip compression and lip stretching. Expressive qualifiers are then used to modulate the expressivity of lip movement. Our model of lip movement is applied on a 3D facial model compliant with MPEG-4 standard. Copyright (C) 2004 John Wiley Sons, Ltd. |
Year | DOI | Venue |
---|---|---|
2004 | 10.1002/cav.32 | COMPUTER ANIMATION AND VIRTUAL WORLDS |
Keywords | Field | DocType |
audio-visual speech, coarticulation, emotion | Viseme,Computer science,Embodied cognition,Coarticulation,Speech recognition,Facial expression,Artificial intelligence,Natural language processing,Dialog system,Sensation,Expressivity | Journal |
Volume | Issue | ISSN |
15 | 3-4 | 1546-4261 |
Citations | PageRank | References |
20 | 1.10 | 9 |
Authors | ||
2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Elisabetta Bevacqua | 1 | 337 | 28.51 |
Catherine Pelachaud | 2 | 2706 | 279.88 |