Channel and frequency attention module for diverse animal sound classification

K. O. Kyungdeuk, Jaihyun Park, David K. Han, Hanseok Ko

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

In-class species classification based on animal sounds is a highly challenging task even with the latest deep learning technique applied. The difficulty of distinguishing the species is further compounded when the number of species is large within the same class. This paper presents a novel approach for fine categorization of animal species based on their sounds by using pre-trained CNNs and a new self-attention module well-suited for acoustic signals The proposed method is shown effective as it achieves average species accuracy of 98.37% and the minimum species accuracy of 94.38%, the highest among the competing baselines, which include CNN’s without self-attention and CNN’s with CBAM, FAM, and CFAM but without pre-training.

Original languageEnglish
Pages (from-to)2615-2618
Number of pages4
JournalIEICE Transactions on Information and Systems
VolumeE102D
Issue number12
DOIs
Publication statusPublished - 2019 Jan 1

Keywords

  • Acoustic signal
  • Artificial intelligence
  • CNN
  • Deep learning
  • Self-attention

ASJC Scopus subject areas

  • Software
  • Hardware and Architecture
  • Computer Vision and Pattern Recognition
  • Electrical and Electronic Engineering
  • Artificial Intelligence

Fingerprint Dive into the research topics of 'Channel and frequency attention module for diverse animal sound classification'. Together they form a unique fingerprint.

Cite this