Music structure analysis using self-similarity matrix and two-stage categorization

Sanghoon Jun, Seungmin Rho, Een Jun Hwang

Research output: Contribution to journalArticle

5 Citations (Scopus)

Abstract

Music tends to have a distinct structure consisting of repetition and variation of components such as verse and chorus. Understanding such a music structure and its pattern has become increasingly important for music information retrieval (MIR). Thus far, many different methods for music segmentation and structure analysis have been proposed; however, each method has its advantages and disadvantages. By considering the significant variations in timbre, articulation and tempo of music, this is still a challenging task. In this paper, we propose a novel method for music segmentation and its structure analysis. For this, we first extract the timbre feature from the acoustic music signal and construct a self-similarity matrix that shows the similarities among the features within the music clip. Further, we determine the candidate boundaries for music segmentation by tracking the standard deviation in the matrix. Furthermore, we perform two-stage categorization: (i) categorization of the segments in a music clip on the basis of the timbre feature and (ii) categorization of segments in the same category on the basis of the successive chromagram features. In this way, each music clip is represented by a sequence of states where each state represents a certain category defined by two-stage categorization. We show the performance of our proposed method through experiments.

Original languageEnglish
Pages (from-to)287-302
Number of pages16
JournalMultimedia Tools and Applications
Volume74
Issue number1
DOIs
Publication statusPublished - 2013 Jan 1

Fingerprint

Information retrieval
Acoustics
Experiments

Keywords

  • Music segmentation
  • Music structure
  • Self-similarity matrix
  • Signal processing

ASJC Scopus subject areas

  • Media Technology
  • Hardware and Architecture
  • Computer Networks and Communications
  • Software

Cite this

Music structure analysis using self-similarity matrix and two-stage categorization. / Jun, Sanghoon; Rho, Seungmin; Hwang, Een Jun.

In: Multimedia Tools and Applications, Vol. 74, No. 1, 01.01.2013, p. 287-302.

Research output: Contribution to journalArticle

@article{30b8ce076bd4445b882d508b03ce2a9a,
title = "Music structure analysis using self-similarity matrix and two-stage categorization",
abstract = "Music tends to have a distinct structure consisting of repetition and variation of components such as verse and chorus. Understanding such a music structure and its pattern has become increasingly important for music information retrieval (MIR). Thus far, many different methods for music segmentation and structure analysis have been proposed; however, each method has its advantages and disadvantages. By considering the significant variations in timbre, articulation and tempo of music, this is still a challenging task. In this paper, we propose a novel method for music segmentation and its structure analysis. For this, we first extract the timbre feature from the acoustic music signal and construct a self-similarity matrix that shows the similarities among the features within the music clip. Further, we determine the candidate boundaries for music segmentation by tracking the standard deviation in the matrix. Furthermore, we perform two-stage categorization: (i) categorization of the segments in a music clip on the basis of the timbre feature and (ii) categorization of segments in the same category on the basis of the successive chromagram features. In this way, each music clip is represented by a sequence of states where each state represents a certain category defined by two-stage categorization. We show the performance of our proposed method through experiments.",
keywords = "Music segmentation, Music structure, Self-similarity matrix, Signal processing",
author = "Sanghoon Jun and Seungmin Rho and Hwang, {Een Jun}",
year = "2013",
month = "1",
day = "1",
doi = "10.1007/s11042-013-1761-9",
language = "English",
volume = "74",
pages = "287--302",
journal = "Multimedia Tools and Applications",
issn = "1380-7501",
publisher = "Springer Netherlands",
number = "1",

}

TY - JOUR

T1 - Music structure analysis using self-similarity matrix and two-stage categorization

AU - Jun, Sanghoon

AU - Rho, Seungmin

AU - Hwang, Een Jun

PY - 2013/1/1

Y1 - 2013/1/1

N2 - Music tends to have a distinct structure consisting of repetition and variation of components such as verse and chorus. Understanding such a music structure and its pattern has become increasingly important for music information retrieval (MIR). Thus far, many different methods for music segmentation and structure analysis have been proposed; however, each method has its advantages and disadvantages. By considering the significant variations in timbre, articulation and tempo of music, this is still a challenging task. In this paper, we propose a novel method for music segmentation and its structure analysis. For this, we first extract the timbre feature from the acoustic music signal and construct a self-similarity matrix that shows the similarities among the features within the music clip. Further, we determine the candidate boundaries for music segmentation by tracking the standard deviation in the matrix. Furthermore, we perform two-stage categorization: (i) categorization of the segments in a music clip on the basis of the timbre feature and (ii) categorization of segments in the same category on the basis of the successive chromagram features. In this way, each music clip is represented by a sequence of states where each state represents a certain category defined by two-stage categorization. We show the performance of our proposed method through experiments.

AB - Music tends to have a distinct structure consisting of repetition and variation of components such as verse and chorus. Understanding such a music structure and its pattern has become increasingly important for music information retrieval (MIR). Thus far, many different methods for music segmentation and structure analysis have been proposed; however, each method has its advantages and disadvantages. By considering the significant variations in timbre, articulation and tempo of music, this is still a challenging task. In this paper, we propose a novel method for music segmentation and its structure analysis. For this, we first extract the timbre feature from the acoustic music signal and construct a self-similarity matrix that shows the similarities among the features within the music clip. Further, we determine the candidate boundaries for music segmentation by tracking the standard deviation in the matrix. Furthermore, we perform two-stage categorization: (i) categorization of the segments in a music clip on the basis of the timbre feature and (ii) categorization of segments in the same category on the basis of the successive chromagram features. In this way, each music clip is represented by a sequence of states where each state represents a certain category defined by two-stage categorization. We show the performance of our proposed method through experiments.

KW - Music segmentation

KW - Music structure

KW - Self-similarity matrix

KW - Signal processing

UR - http://www.scopus.com/inward/record.url?scp=84921701391&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84921701391&partnerID=8YFLogxK

U2 - 10.1007/s11042-013-1761-9

DO - 10.1007/s11042-013-1761-9

M3 - Article

AN - SCOPUS:84921701391

VL - 74

SP - 287

EP - 302

JO - Multimedia Tools and Applications

JF - Multimedia Tools and Applications

SN - 1380-7501

IS - 1

ER -