Abstract | ||
---|---|---|
Traditional approaches to automatic emotion recognition from speech typically make use of utterance level prosodic features. Still, a great deal of useful information about expressivity and emotion can be gained from segmental spectral features, which provide a more detailed description of the speech signal, or from measurements from specific regions of the utterance, such as the stressed vowels. Here we introduce a novel set of spectral features for emotion recognition: statistics of Mel-Frequency Spectral Coefficients computed over three phoneme type classes of interest: stressed vowels, unstressed vowels and consonants in the utterance. We investigate performance of our features in the task of speaker-independent emotion recognition using two publicly available datasets. Our experimental results clearly indicate that indeed both the richer set of spectral features and the differentiation between phoneme type classes are beneficial for the task. Classification accuracies are consistently higher for our features compared to prosodic features or utterance-level spectral features. Combination of our phoneme class features with prosodic features leads to even further improvement. |
Year | Venue | Keywords |
---|---|---|
2009 | INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5 | emotion recognition |
Field | DocType | Citations |
Emotion recognition,Computer science,Utterance,Speech recognition,Natural language processing,Artificial intelligence,Expressivity | Conference | 3 |
PageRank | References | Authors |
0.42 | 15 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Dmitri Bitouk | 1 | 238 | 9.65 |
Ani Nenkova | 2 | 1831 | 109.14 |
Ragini Verma | 3 | 731 | 46.37 |