Abstract | ||
---|---|---|
Finding an appropriate feature representation for audio data is central to speech emotion recognition. Most existing audio features rely on hand-crafted feature encoding techniques, such as the AVEC challenge feature set. An alternative approach is to use features that are learned automatically. This has the advantage of generalizing well to new data, particularly if the features are learned in an unsupervised manner with less restrictions on the data itself. In this work, we adopt the sparse coding framework as a means to automatically represent features from audio and propose a hierarchical sparse coding (HSC) scheme. Experimental results indicate that the obtained features, in an unsupervised fashion, are able to capture useful properties of the speech that distinguish between emotions. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1016/j.specom.2018.01.006 | Speech Communication |
Keywords | Field | DocType |
Affective computing,Speech emotion recognition,Sparse coding,Support vector regression | Pattern recognition,Computer science,Emotion recognition,Neural coding,Generalization,Speech recognition,Feature set,Artificial intelligence,Encoding (memory) | Journal |
Volume | ISSN | Citations |
99 | 0167-6393 | 3 |
PageRank | References | Authors |
0.36 | 43 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Diana Torres-Boza | 1 | 3 | 0.36 |
Meshia Cédric Oveneke | 2 | 28 | 7.39 |
Fengna Wang | 3 | 41 | 2.94 |
Dongmei Jiang | 4 | 5 | 1.74 |
Werner Verhelst | 5 | 431 | 51.55 |
Hichem Sahli | 6 | 475 | 65.19 |