@inproceedings{ee8813ffd1c8446b8fbe05736be26966,
title = "MANNER: MULTI-VIEW ATTENTION NETWORK FOR NOISE ERASURE",
abstract = "In the field of speech enhancement, time domain methods have difficulties in achieving both high performance and efficiency. Recently, dual-path models have been adopted to represent long sequential features, but they still have limited representations and poor memory efficiency. In this study, we propose Multi-view Attention Network for Noise ERasure (MANNER) consisting of a convolutional encoder-decoder with a multi-view attention block, applied to the time-domain signals. MANNER efficiently extracts three different representations from noisy speech and estimates high-quality clean speech. We evaluated MANNER on the VoiceBank-DEMAND dataset in terms of five objective speech quality metrics. Experimental results show that MANNER achieves state-of-the-art performance while efficiently processing noisy speech.",
keywords = "multi-view attention, speech enhancement, time domain, u-net",
author = "Park, {Hyun Joon} and Kang, {Byung Ha} and Wooseok Shin and Kim, {Jin Sob} and Han, {Sung Won}",
note = "Funding Information: This research was supported by Brain Korea 21 FOUR. This research was also supported by Korea University Grant (K2107521) and a Korea Tech-noComplex Foundation Grant (R2112651). Funding Information: This research was supported by Brain Korea 21 FOUR. This research was also supported by Korea University Grant (K2107521) and a Korea TechnoComplex Foundation Grant (R2112651). Publisher Copyright: {\textcopyright} 2022 IEEE; 47th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022 ; Conference date: 23-05-2022 Through 27-05-2022",
year = "2022",
doi = "10.1109/ICASSP43922.2022.9747120",
language = "English",
series = "ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "7842--7846",
booktitle = "2022 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022 - Proceedings",
}