Title
Tempo and beat tracking for audio signals with music genre classification
Abstract
Most people follow the music to hum or the rhythm to tap sometimes. We may get different meanings of a music style if it is explained or felt by different people. Therefore we cannot obtain a very explicit answer if there is no music notation. Tempo and beats are very important elements in the perceptual music. Therefore, tempo estimation and beat tracking are fundamental techniques in automatic audio processing, which are crucial to multimedia applications. We first develop an artificial neural network to classify the music excerpts into the evaluation preference. And then, with the preference classification, we can obtain accurate estimation for tempo and beats, by either Ellis's method or Dixon's method. We test our method with mixed data set which contains ten music genres from the 'ballroom dancer' database. Our experimental results show that the accuracy of our method is higher than only one individual Ellis's method or Dixon's method.
Year
DOI
Venue
2009
10.1504/IJIIDS.2009.027687
IJIIDS
Keywords
Field
DocType
classification.,different meaning,beat,neural network,music style,evaluation preference,tempo,music genre classification,music genre,different people,music notation,perceptual music,tempo estimation,music excerpt,audio signal,audio processing,accurate estimation,artificial neural network,classification,artificial neural networks,audio signals,multimedia
Audio signal,Musical notation,Computer science,Speech recognition,Beat (music),Audio signal processing,Artificial neural network,Rhythm,Perception,Sound recording and reproduction
Journal
Volume
Issue
Citations 
3
3
2
PageRank 
References 
Authors
0.42
4
3
Name
Order
Citations
PageRank
Mao-Yuan Kao1101.01
Chang-Biau Yang220326.74
Shyue-Horng Shiau3203.00