201 122

Full metadata record

DC FieldValueLanguage
dc.contributor.author장준혁-
dc.date.accessioned2022-03-29T02:35:01Z-
dc.date.available2022-03-29T02:35:01Z-
dc.date.issued2020-07-
dc.identifier.citationAPPLIED SCIENCES-BASEL, v. 10, no. 13, article no. 4602en_US
dc.identifier.issn2076-3417-
dc.identifier.urihttps://www.mdpi.com/2076-3417/10/13/4602-
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/169489-
dc.description.abstractSpeech recognition for intelligent robots seems to suffer from performance degradation due to ego-noise. The ego-noise is caused by the motors, fans, and mechanical parts inside the intelligent robots especially when the robot moves or shakes its body. To overcome the problems caused by the ego-noise, we propose a robust speech recognition algorithm that uses motor-state information of the robot as an auxiliary feature. For this, we use two deep neural networks (DNN) in this paper. Firstly, we design the latent features using a bottleneck layer, one of the internal layers having a smaller number of hidden units relative to the other layers, to represent whether the motor is operating or not. The latent features maximizing the representation of the motor-state information are generated by taking the motor data and acoustic features as the input of the first DNN. Secondly, once the motor-state dependent latent features are designed at the first DNN, the second DNN, accounting for acoustic modeling, receives the latent features as the input along with the acoustic features. We evaluated the proposed system on LibriSpeech database. The proposed network enables efficient compression of the acoustic and motor-state information, and the resulting word error rate (WER) are superior to that of a conventional speech recognition system.en_US
dc.description.sponsorshipThis work was supported by the research fund of Signal Intelligence Research Center supervised by the Defense Acquisition Program Administration and the Agency for Defense Development of Korea.en_US
dc.language.isoenen_US
dc.publisherMDPIen_US
dc.subjectautomatic speech recognitionen_US
dc.subjecthuman-robot interactionen_US
dc.subjectdeep learningen_US
dc.subjectbottleneck layeren_US
dc.subjectlatent featureen_US
dc.subjectbottleneck networken_US
dc.titleAugmented Latent Features of Deep Neural Network-Based Automatic Speech Recognition for Motor-Driven Robotsen_US
dc.typeArticleen_US
dc.relation.no13-
dc.relation.volume10-
dc.identifier.doi10.3390/app10134602-
dc.relation.page1-10-
dc.relation.journalAPPLIED SCIENCES-BASEL-
dc.contributor.googleauthorLee, Moa-
dc.contributor.googleauthorChang, Joon-Hyuk-
dc.relation.code2020047168-
dc.sector.campusS-
dc.sector.daehakCOLLEGE OF ENGINEERING[S]-
dc.sector.departmentSCHOOL OF ELECTRONIC ENGINEERING-
dc.identifier.pidjchang-
dc.identifier.orcidhttps://orcid.org/0000-0003-2610-2323-


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE