Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | 신현철 | - |
dc.date.accessioned | 2021-12-23T04:08:04Z | - |
dc.date.available | 2021-12-23T04:08:04Z | - |
dc.date.issued | 2021-02 | - |
dc.identifier.citation | MULTIDIMENSIONAL SYSTEMS AND SIGNAL PROCESSING, v. 32, No. 3, Page. 897-913 | en_US |
dc.identifier.issn | 0923-6082 | - |
dc.identifier.issn | 1573-0824 | - |
dc.identifier.uri | https://link.springer.com/article/10.1007%2Fs11045-021-00764-1 | - |
dc.identifier.uri | https://repository.hanyang.ac.kr/handle/20.500.11754/166986 | - |
dc.description.abstract | Detecting small-scale pedestrians in aerial images is a challenging task that can be difficult even for humans. Observing that the single image based method cannot achieve robust performance because of the poor visual cues of small instances. Considering that multiple frames may provide more information to detect such difficult case instead of only single frame, we design a novel video based pedestrian detection method with a two-stream network pipeline to fully utilize the temporal and contextual information of a video. An aggregated feature map is proposed to absorb the spatial and temporal information with the help of spatial and temporal sub-networks. To better capture motion information, a more refined flow net (SPyNet) is adopted instead of a simple flownet. In the spatial stream subnetwork, we modified the backbone network structure by increasing the feature map resolution with relatively larger receptive field to make it suitable for small-scale detection. Experimental results based on drone video datasets demonstrate that our approach improves detection accuracy in the case of small-scale instances and reduces false positive detections. By exploiting the temporal information and aggregating the feature maps, our two-stream method improves the detection performance by 8.48% in mean Average Precision (mAP) from that of the basic single stream R-FCN method, and it outperforms the state-of-the-art method by 3.09% on the Okutama Human-action dataset. | en_US |
dc.language.iso | en_US | en_US |
dc.publisher | SPRINGER | en_US |
dc.subject | Pedestrian detection | en_US |
dc.subject | Feature aggregation | en_US |
dc.subject | Drone vision | en_US |
dc.subject | Neural network | en_US |
dc.subject | Deep learning | en_US |
dc.title | Two-stream small-scale pedestrian detection network with feature aggregation for drone-view videos | en_US |
dc.type | Article | en_US |
dc.relation.no | 3 | - |
dc.relation.volume | 32 | - |
dc.identifier.doi | 10.1007/s11045-021-00764-1 | - |
dc.relation.page | 897-913 | - |
dc.relation.journal | MULTIDIMENSIONAL SYSTEMS AND SIGNAL PROCESSING | - |
dc.contributor.googleauthor | Xie, Han | - |
dc.contributor.googleauthor | Shin, Hyunchul | - |
dc.relation.code | 2021006555 | - |
dc.sector.campus | E | - |
dc.sector.daehak | COLLEGE OF ENGINEERING SCIENCES[E] | - |
dc.sector.department | DIVISION OF ELECTRICAL ENGINEERING | - |
dc.identifier.pid | shin | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.