161 0

Full metadata record

DC FieldValueLanguage
dc.contributor.author최준원-
dc.date.accessioned2022-10-17T05:48:28Z-
dc.date.available2022-10-17T05:48:28Z-
dc.date.issued2021-01-
dc.identifier.citation25th International Conference on Pattern Recognition (ICPR), page. 4505-4512en_US
dc.identifier.issn1051-4651en_US
dc.identifier.urihttps://ieeexplore.ieee.org/document/9412795en_US
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/175472-
dc.description.abstractConvolutional neural networks (CNNs) have led us to achieve significant progress in object detection research. To detect objects of various sizes, object detectors often exploit the hierarchy of the multiscale feature maps called feature pyramids, which are readily obtained by the CNN architecture. However, the performance of these object detectors is limited because the bottom-level feature maps, which experience fewer convolutional layers, lack the semantic information needed to capture the characteristics of the small objects. To address such problems, various methods have been proposed to increase the depth for the bottom-level features used for object detection. While most approaches are based on the generation of additional features through the top-down pathway with lateral connections, our approach directly fuses multi-scale feature maps using bidirectional long short-term memory (biLSTM) in an effort to leverage the gating functions and parameter-sharing in generating deeply fused semantics. The resulting semantic information is redistributed to the individual pyramidal feature at each scale through the channel-wise attention model. We integrate our semantic combining and attentive redistribution feature network (ScarfNet) with the baseline object detectors, i.e., Faster R-CNN, single-shot multibox detector (SSD), and RetinaNet. Experimental results show that our method offers a significant performance gain over the baseline detectors and outperforms the competing multiscale fusion methods in the PASCAL VOC and COCO detection benchmarks.en_US
dc.description.sponsorshipThis work was supported in part by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2020-0-01373, Artificial Intelligence Graduate School Program (Hanyang University)) and the National Research Foundation of Korea (NRF) grant funded by the Korea government(MSIT) (No. 2020R1A2C2012146).en_US
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.titleScarfNet: Multi-scale Features with Deeply Fused and Redistributed Semantics for Enhanced Object Detectionen_US
dc.typeArticleen_US
dc.identifier.doi10.1109/ICPR48806.2021.9412795en_US
dc.relation.page4505-4512-
dc.contributor.googleauthorYoo, Jin Hyeok-
dc.contributor.googleauthorKum, Dongsuk-
dc.contributor.googleauthorChoi, Jun Won-
dc.relation.code20210142-
dc.sector.campusS-
dc.sector.daehakCOLLEGE OF ENGINEERING[S]-
dc.sector.departmentSCHOOL OF ELECTRICAL AND BIOMEDICAL ENGINEERING-
dc.identifier.pidjunwchoi-
Appears in Collections:
COLLEGE OF ENGINEERING[S](공과대학) > ELECTRICAL AND BIOMEDICAL ENGINEERING(전기·생체공학부) > Articles
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE