Music structure analysis using self-similarity matrix and two-stage categorization

Sanghoon Jun, Seungmin Rho, Een Jun Hwang

Research output: Contribution to journalArticle

5 Citations (Scopus)

Abstract

Music tends to have a distinct structure consisting of repetition and variation of components such as verse and chorus. Understanding such a music structure and its pattern has become increasingly important for music information retrieval (MIR). Thus far, many different methods for music segmentation and structure analysis have been proposed; however, each method has its advantages and disadvantages. By considering the significant variations in timbre, articulation and tempo of music, this is still a challenging task. In this paper, we propose a novel method for music segmentation and its structure analysis. For this, we first extract the timbre feature from the acoustic music signal and construct a self-similarity matrix that shows the similarities among the features within the music clip. Further, we determine the candidate boundaries for music segmentation by tracking the standard deviation in the matrix. Furthermore, we perform two-stage categorization: (i) categorization of the segments in a music clip on the basis of the timbre feature and (ii) categorization of segments in the same category on the basis of the successive chromagram features. In this way, each music clip is represented by a sequence of states where each state represents a certain category defined by two-stage categorization. We show the performance of our proposed method through experiments.

Original languageEnglish
Pages (from-to)287-302
Number of pages16
JournalMultimedia Tools and Applications
Volume74
Issue number1
DOIs
Publication statusPublished - 2013 Jan 1

    Fingerprint

Keywords

  • Music segmentation
  • Music structure
  • Self-similarity matrix
  • Signal processing

ASJC Scopus subject areas

  • Media Technology
  • Hardware and Architecture
  • Computer Networks and Communications
  • Software

Cite this