Dual-layer bag-of-frames model for music genre classification
Journal
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
ISBN
9781479903566
Date Issued
2013-10-18
Author(s)
Abstract
This paper concerns the development of a music dictionary-based model for summarizing local feature descriptors computed over time. Comparing to a holistic representation, this text-like, bag-of-frames representation better captures the rich and time-varying information of music. However, the dictionary used in classical bag-of-frames model only captures frame-level elements of the music; thus, there exists a semantic gap between the dictionary element and commonly seen music description. In order to reduce the gap, a new feature representation called dual-layer bag-of-frames is proposed in this paper. It models the music with a two layer structure, where the first-layer dictionary captures the frame-level characteristics, and the second-layer dictionary captures the segment-level semantics. This hierarchical structure resembles the alphabet-word-document structure of text. Our result demonstrates that the proposed dual-layer bag-of-frames feature achieves state-of-the-art accuracy of music genre classification. The classification accuracy for the GTZAN benchmark reaches 86.7% with dictionary trained from GTZAN, and 83.6% with dictionary trained from another data set USPOP. © 2013 IEEE.
Subjects
audio alphabets | audio words | deep structure | music genre classification | Sparse coding
SDGs
Type
conference paper
