Abstract | ||
---|---|---|
Music is highly related to human affective feelings with different kinds of emotions may be embedded in a music work simultaneously. Hence, how to extract emotions from music has been a hot topic for music information retrieval over the past few decades. To this end, a considerable number of multi-labeling studies have been conducted on tagging music emotions. In this paper, we conduct a comparative analysis of state-of-the-art methods for music emotion annotation through extensive experimental evaluations. Comparative experiments were performed on real dataset CAL500 with different evaluation metrics. Moreover, to reveal the robustness, the compared algorithms including different domains of annotation ones were examined with simple and complex types of emotions. The experimental results provide the researchers with insightful ideas in algorithm design for emotionalizing music from technical point of view. |
Year | DOI | Venue |
---|---|---|
2013 | 10.1109/ICMEW.2013.6618344 | ICME Workshops |
Keywords | Field | DocType |
human affective feelings,multi-labeling,tagging,music,information retrieval,annotation,empirical analysis,multilabeling algorithms,emotion recognition,music emotion,music emotion annotation,frequency modulation,semantics,labeling,indexes,probabilistic logic | Music information retrieval,Algorithm design,Annotation,Computer science,Emotion recognition,Algorithm,Robustness (computer science),Natural language processing,Artificial intelligence,Feeling | Conference |
ISSN | Citations | PageRank |
2330-7927 | 1 | 0.37 |
References | Authors | |
10 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ja-Hwung Su | 1 | 329 | 24.53 |
Yi-Cheng Tsai | 2 | 1 | 0.71 |
Vincent S. Tseng | 3 | 2923 | 161.33 |