Full metadata record

DC FieldValueLanguage
dc.contributor.author임종우-
dc.date.accessioned2018-11-05T02:09:31Z-
dc.date.available2018-11-05T02:09:31Z-
dc.date.issued2016-09-
dc.identifier.citationComputer Vision - 14th European Conference, ECCV 2016, Proceedings. (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics, 2016), v. 9909, Page. 415-433en_US
dc.identifier.isbn978-3-31-946453-4-
dc.identifier.isbn978-3-319-46454-1-
dc.identifier.issn1611-3349-
dc.identifier.issn0302-9743-
dc.identifier.urihttps://link.springer.com/chapter/10.1007%2F978-3-319-46454-1_26-
dc.identifier.urihttps://repository.hanyang.ac.kr/handle/20.500.11754/79117-
dc.description.abstractMulti-face tracking in unconstrained videos is a challenging problem as faces of one person often appear drastically different in multiple shots due to significant variations in scale, pose, expression, illumination, and make-up. Low-level features used in existing multi-target tracking methods are not effective for identifying faces with such large appearance variations. In this paper, we tackle this problem by learning discriminative, video-specific face features using convolutional neural networks (CNNs). Unlike existing CNN-based approaches that are only trained on large-scale face image datasets offline, we further adapt the pre-trained face CNN to specific videos using automatically discovered training samples from tracklets. Our network directly optimizes the embedding space so that the Euclidean distances correspond to a measure of semantic face similarity. This is technically realized by minimizing an improved triplet loss function. With the learned discriminative features, we apply the Hungarian algorithm to link tracklets within each shot and the hierarchical clustering algorithm to link tracklets across multiple shots to form final trajectories. We extensively evaluate the proposed algorithm on a set of TV sitcoms and music videos and demonstrate significant performance improvement over existing techniques.en_US
dc.description.sponsorshipThe work is partially supported by National Basic Research Program of China (973 Program, 2015CB351705), NSFC (61332018), Office of Naval Research (N0014-16-1-2314), R&D programs by NRF (2014R1A1A2058501) and MSIP/IITP (IITP-2016-H8601-16-1005) of Korea, NSF CAREER (1149783) and gifts from Adobe and NVIDIA.en_US
dc.language.isoenen_US
dc.publisherECCV/Springeren_US
dc.subjectConvolutional Neural Networken_US
dc.subjectMusic Videoen_US
dc.subjectMultiple Shoten_US
dc.subjectVisual Constrainten_US
dc.subjectData Association Problemen_US
dc.titleTracking Persons-of-Interest via Adaptive Discriminative Featuresen_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1007/978-3-319-46454-1_26-
dc.relation.page415-433-
dc.contributor.googleauthorZhang, Shun-
dc.contributor.googleauthorGong, Yihong-
dc.contributor.googleauthorHuang, Jia-Bin-
dc.contributor.googleauthorLim, Jongwoo-
dc.contributor.googleauthorWang, Jinjun-
dc.contributor.googleauthorAhuja, Narendra-
dc.contributor.googleauthorYang, Ming-Hsuan-
dc.relation.code20160042-
dc.sector.campusS-
dc.sector.daehakCOLLEGE OF ENGINEERING[S]-
dc.sector.departmentDEPARTMENT OF COMPUTER SCIENCE-
dc.identifier.pidjlim-
Appears in Collections:
COLLEGE OF ENGINEERING[S](공과대학) > COMPUTER SCIENCE(컴퓨터소프트웨어학부) > Articles
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML


qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE