2015
DOI: 10.1109/tip.2015.2493448
|View full text |Cite
|
Sign up to set email alerts
|

A Benchmark and Comparative Study of Video-Based Face Recognition on COX Face Database

Abstract: Face recognition with still face images has been widely studied, while the research on video-based face recognition is inadequate relatively, especially in terms of benchmark datasets and comparisons. Real-world video-based face recognition applications require techniques for three distinct scenarios: 1) Videoto-Still (V2S); 2) Still-to-Video (S2V); and 3) Video-to-Video (V2V), respectively, taking video or still image as query or target. To the best of our knowledge, few datasets and evaluation protocols have… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
82
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 136 publications
(83 citation statements)
references
References 42 publications
0
82
0
1
Order By: Relevance
“…Fig 2(a) shows ROIs of 5 selected target individuals and their test video correspondence is recorded with 3 cameras above different portals. COX-S2V [37] consists of 1000 subjects, where each subject has a high quality still images under controlled conditions, and four lower-quality facial trajectories captured under uncontrolled conditions. Each trajectory has 25 faces, where ROIs taken from these videos encounter changes in illumination, expression, scale, viewpoint, and blur.…”
Section: Methodology For Validationmentioning
confidence: 99%
“…Fig 2(a) shows ROIs of 5 selected target individuals and their test video correspondence is recorded with 3 cameras above different portals. COX-S2V [37] consists of 1000 subjects, where each subject has a high quality still images under controlled conditions, and four lower-quality facial trajectories captured under uncontrolled conditions. Each trajectory has 25 faces, where ROIs taken from these videos encounter changes in illumination, expression, scale, viewpoint, and blur.…”
Section: Methodology For Validationmentioning
confidence: 99%
“…In order to evaluate the performance of the proposed S+V model for still-to-video FR, an extensive series of experiments are conducted on Chokepoint 2 [24] and COX-S2V 3 [25] datasets. Chokepoint [24] and COX-S2V [25] datasets are suitable for ex-…”
Section: Datasetsmentioning
confidence: 99%
“…Moreover, since most state-of-the-art FR methods rely on Convolution Neural Network (CNN) architectures such as ResNet [20] and VGGNet [21], the model is fed with CNN features extracted from the atoms of dictionaries [22,23], in order to further improve still-tovideo FR accuracy. Performance of the SRC implementation is evaluated on two public video FR databases -Chokepoint [24] and COX-S2V [25].…”
Section: Introductionmentioning
confidence: 99%
“…There are 293 identities from PaSC testing dataset, including 9376 still images (about 32 images per subject), and 2802 videos (approximately ten videos per person and 100 frames per video). Another dataset is COX [25], which has 1,000 subjects, 3 videos captured for each subject with 3 different camcorders. An interactive tool for annotating facial attributes was developed, which displayed multiple face images from the same subject.…”
Section: A Datamentioning
confidence: 99%
“…The COX [25] consists of 1,000 subjects and three videos for each subject. We focus on the videos, which contain several frames, and demonstrate the attribute inconsistency issue.…”
Section: Videos On Coxmentioning
confidence: 99%