127 47

Full metadata record

DC FieldValueLanguage
dc.contributor.author노정진-
dc.date.accessioned2023-07-13T01:43:12Z-
dc.date.available2023-07-13T01:43:12Z-
dc.date.issued2022-04-
dc.identifier.citationAPPLIED SCIENCES-BASEL, v. 12, NO. 7, article no. 3686, Page. 1-11-
dc.identifier.issn2076-3417;2076-3417-
dc.identifier.urihttps://www.mdpi.com/2076-3417/12/7/3686en_US
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/183419-
dc.description.abstractOwing to its wide range of applications, 3D object detection has attracted increasing attention in computer vision tasks. Most existing 3D object detection methods are based on Lidar point cloud data. However, these methods have some limitations in localization consistency and classification confidence, due to the irregularity and sparsity of Light Detection and Ranging (LiDAR) point cloud data. Inspired by the complementary characteristics of Lidar and camera sensors, we propose a new end-to-end learnable framework named Point-Image Fusion Network (PIFNet) to integrate the LiDAR point cloud and camera images. To resolve the problem of inconsistency in the localization and classification, we designed an Encoder-Decoder Fusion (EDF) module to extract the image features effectively, while maintaining the fine-grained localization information of objects. Furthermore, a new effective fusion module is proposed to integrate the color and texture features from images and the depth information from the point cloud. This module can enhance the irregularity and sparsity problem of the point cloud features by capitalizing the fine-grained information from camera images. In PIFNet, each intermediate feature map is fed into the fusion module to be integrated with its corresponding point-wise features. Furthermore, point-wise features are used instead of voxel-wise features to reduce information loss. Extensive experiments using the KITTI dataset demonstrate the superiority of PIFNet over other state-of-the-art methods. Compared with several state-of-the-art methods, our approach outperformed by 1.97% in mean Average Precision (mAP) and by 2.86% in Average Precision (AP) for the hard cases on the KITTI 3D object detection benchmark.-
dc.description.sponsorshipThis research was funded by the Ministry of Trade, Industry & Energy (MOTIE, Korea) under 343 Industrial Technology Innovation Program (10080619).-
dc.languageen-
dc.publisherMDPI-
dc.subject3D object detection-
dc.subjectlidar point cloud-
dc.subjectcamera images-
dc.subjectobject detection-
dc.titlePIFNet: 3D Object Detection Using Joint Image and Point Cloud Features for Autonomous Driving-
dc.typeArticle-
dc.relation.no7-
dc.relation.volume12-
dc.identifier.doi10.3390/app12073686-
dc.relation.page1-11-
dc.relation.journalAPPLIED SCIENCES-BASEL-
dc.contributor.googleauthorZheng, Wenqi-
dc.contributor.googleauthorXie, Han-
dc.contributor.googleauthorChen, Yunfan-
dc.contributor.googleauthorRoh, Jeongjin-
dc.contributor.googleauthorShin, Hyunchul-
dc.sector.campusE-
dc.sector.daehak공학대학-
dc.sector.department전자공학부-
dc.identifier.pidjroh-
dc.identifier.article3686-


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE