Full metadata record

DC FieldValueLanguage
dc.contributor.author김기범-
dc.date.accessioned2024-04-25T00:47:37Z-
dc.date.available2024-04-25T00:47:37Z-
dc.date.issued2022-09-20-
dc.identifier.citationMULTIMEDIA TOOLS AND APPLICATIONS, v. 82, NO 9, Page. 13401-13430en_US
dc.identifier.issn1380-7501en_US
dc.identifier.issn1573-7721en_US
dc.identifier.urihttps://information.hanyang.ac.kr/#/eds/detail?an=edssjs.86A3EB14&dbId=edssjsen_US
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/190002-
dc.description.abstractRecent advances in visionary technologies impacted multi-object recognition and scene understanding. Such scene-understanding tasks are a demanding part of several technologies such as augmented reality based scene integration, robotic navigation, autonomous driving and tourist guide applications. Incorporating visual information in contextually unified segments, super-pixel-based approaches significantly mitigate the clutter, which is normal in pixel wise frameworks during scene understanding. Super-pixels allow customized shapes and variable size patches of connected components to be obtained. Furthermore, the computational time for these segmentation approaches can significantly decreased due to the reduced number of super-pixel target clusters. Hence, the super pixel-based approaches are more commonly used in robotics, computer vision and other intelligent systems. In this paper, we propose a Maximum Entropy scaled Super-Pixels (MEsSP) Segmentation method that encapsulates super-pixel segmentation based on an Entropy Model and utilizes local energy terms to label the pixels. Initially, after acquisition and pre-processing, image is segmented by two different methods: Fuzzy C-Means (FCM) and MEsSP. Then, to extract the features from these segmented objects, the dynamic geometrical features, fast Fourier transform (FFT), blob extraction, Maximally Stable Extremal Regions (MSER) and KAZE features are extracted using the bag of features approach. Then, to categorize the objects, multiple kernel learning is applied. Finally, a deep belief network (DBN) assigns the relevant labels to the scenes based on the categorized objects, intersection over union scores and dice similarity coefficient. The experimental results regarding multiple objects recognition accuracy, precision, recall and F1 scores over PASCAL VOC, Caltech 101 and UIUC Sports datasets show a remarkable performance. In addition, the evaluation of proposed scene recognition method over these benchmark datasets outperforms the state of the art (SOTA) methods.en_US
dc.description.sponsorshipThis research was supported by the Ministry of Culture, Sports and Tourism and Korea Creative Content Agency (Project Number: R2021040093).en_US
dc.languageen_USen_US
dc.publisherSPRINGERen_US
dc.relation.ispartofseriesv. 82, NO 9;13401-13430-
dc.subjectBag of featuresen_US
dc.subjectDeep belief networken_US
dc.subjectEntropy-scaled segmentationen_US
dc.subjectSuper-pixelsen_US
dc.titleMaximum entropy scaled super pixels segmentation for multi-object detection and scene recognition via deep belief networken_US
dc.typeArticleen_US
dc.relation.no9-
dc.relation.volume82-
dc.identifier.doi10.1007/s11042-022-13717-yen_US
dc.relation.page13401-13430-
dc.relation.journalMULTIMEDIA TOOLS AND APPLICATIONS-
dc.contributor.googleauthorRafique, Adnan Ahmed-
dc.contributor.googleauthorGochoo, Munkhjargal-
dc.contributor.googleauthorJalal, Ahmad-
dc.contributor.googleauthorKim, Kibum-
dc.relation.code2023037732-
dc.sector.campusE-
dc.sector.daehakCOLLEGE OF COMPUTING[E]-
dc.sector.departmentSCHOOL OF MEDIA, CULTURE, AND DESIGN TECHNOLOGY-
dc.identifier.pidkibum-
Appears in Collections:
COLLEGE OF COMPUTING[E](소프트웨어융합대학) > MEDIA, CULTURE, AND DESIGN TECHNOLOGY(ICT융합학부) > Articles
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE