DocumentCode :
3697409
Title :
Modeling musical rhythmatscale with the music Genome project
Author :
Matthew Prockup;Andreas F. Ehmann;Fabien Gouyon;Erik M. Schmidt;Youngmoo E. Kim
Author_Institution :
Drexel University, ECE Dept., 3141 Chestnut Street, Philadelphia, PA 19104
fYear :
2015
Firstpage :
1
Lastpage :
5
Abstract :
Musical meter and attributes of the rhythmic feel such as swing, syncopation, and danceability are crucial when defining musical style. However, they have attracted relatively little attention from the Music Information Retrieval (MIR) community and, when addressed, have proven difficult to model from music audio signals. In this paper, we propose a number of audio features for modeling meter and rhythmic feel. These features are first evaluated and compared to timbral features in the common task of ballroom genre classification. These features are then used to learn individual models for a total of nine rhythmic attributes covering meter and feel using an industrial-sized corpus of over one million examples labeled by experts from Pandora® Internet Radio´s Music Genome Project®. Linear models are shown to be powerful, representing these attributes with high accuracy at scale.
Keywords :
"Rhythm","Transforms","Multiple signal classification","Context","Genomics","Bioinformatics"
Publisher :
ieee
Conference_Titel :
Applications of Signal Processing to Audio and Acoustics (WASPAA), 2015 IEEE Workshop on
Type :
conf
DOI :
10.1109/WASPAA.2015.7336891
Filename :
7336891
Link To Document :
بازگشت