305 93

Full metadata record

DC FieldValueLanguage
dc.contributor.author장준혁-
dc.date.accessioned2021-05-13T02:09:53Z-
dc.date.available2021-05-13T02:09:53Z-
dc.date.issued2020-03-
dc.identifier.citationSENSORS, v. 20, no. 7, article no. 1883en_US
dc.identifier.issn1424-8220-
dc.identifier.urihttps://www.mdpi.com/1424-8220/20/7/1883-
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/162020-
dc.description.abstractIn this paper, we propose joint optimization of deep neural network (DNN)-supported dereverberation and beamforming for the convolutional recurrent neural network (CRNN)-based sound event detection (SED) in multi-channel environments. First, the short-time Fourier transform (STFT) coefficients are calculated from multi-channel audio signals under the noisy and reverberant environments, which are then enhanced by the DNN-supported weighted prediction error (WPE) dereverberation with the estimated masks. Next, the STFT coefficients of the dereverberated multi-channel audio signals are conveyed to the DNN-supported minimum variance distortionless response (MVDR) beamformer in which DNN-supported MVDR beamforming is carried out with the source and noise masks estimated by the DNN. As a result, the single-channel enhanced STFT coefficients are shown at the output and tossed to the CRNN-based SED system, and then, the three modules are jointly trained by the single loss function designed for SED. Furthermore, to ease the difficulty of training a deep learning model for SED caused by the imbalance in the amount of data for each class, the focal loss is used as a loss function. Experimental results show that joint training of DNN-supported dereverberation and beamforming with the SED model under the supervision of focal loss significantly improves the performance under the noisy and reverberant environments.en_US
dc.description.sponsorshipThis research was supported by Projects for Research and Development of Police science and Technology under Center for Research and Development of Police science and Technology and Korean National Police Agency funded by the Ministry of Science and ICT (PA-J000001-2017-101).en_US
dc.language.isoenen_US
dc.publisherMDPIen_US
dc.subjectsound event detectionen_US
dc.subjectdereverberationen_US
dc.subjectacoustic beamformingen_US
dc.subjectconvolutional recurrent neural networken_US
dc.subjectjoint optimizationen_US
dc.titleJoint Optimization of Deep Neural Network-Based Dereverberation and Beamforming for Sound Event Detection in Multi-Channel Environmentsen_US
dc.typeArticleen_US
dc.relation.no7-
dc.relation.volume20-
dc.identifier.doi10.3390/s20071883-
dc.relation.page1-13-
dc.relation.journalSENSORS-
dc.contributor.googleauthorNoh, Kyoungjin-
dc.contributor.googleauthorChang, Joon-Hyuk-
dc.relation.code2020053568-
dc.sector.campusS-
dc.sector.daehakCOLLEGE OF ENGINEERING[S]-
dc.sector.departmentDEPARTMENT OF ELECTRONIC ENGINEERING-
dc.identifier.pidjchang-
dc.identifier.orcidhttps://orcid.org/0000-0003-2610-2323-


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE