2023
DOI: 10.1109/tpami.2022.3169734
|View full text |Cite
|
Sign up to set email alerts
|

WebFace260M: A Benchmark for Million-Scale Deep Face Recognition

Abstract: Face benchmarks empower the research community to train and evaluate high-performance face recognition systems. In this paper, we contribute a new million-scale recognition benchmark, containing uncurated 4M identities/260M faces (WebFace260M) and cleaned 2M identities/42M faces (WebFace42M) training data, as well as an elaborately designed time-constrained evaluation protocol. Firstly, we collect 4M name lists and download 260M faces from the Internet. Then, a Cleaning Automatically utilizing Self-Training (C… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(6 citation statements)
references
References 114 publications
0
6
0
Order By: Relevance
“…Training the model on much larger datasets before getting applied in production will certainly perform even better, while no training is necessary when just applying the model to new test users. This is similar to face recognition, whose models are used in many photo applications today and do not need any on-device re-training, since they were trained on millions of individuals' faces [45].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Training the model on much larger datasets before getting applied in production will certainly perform even better, while no training is necessary when just applying the model to new test users. This is similar to face recognition, whose models are used in many photo applications today and do not need any on-device re-training, since they were trained on millions of individuals' faces [45].…”
Section: Discussionmentioning
confidence: 99%
“…We also observe that the performance is not saturated for the maximally available training users for our dataset. This suggests that additional users during training can substantially improve current results, which is a common observation in related applications such as face recognition, where production-grade models are typically trained on millions of subjects [45]. This motivates the collection of larger datasets for the task of VR user identification.…”
Section: Is the Model Robust To Different Seeds?mentioning
confidence: 98%
“…Considering that AdaFace improves largely the FR performance on low quality academic datasets [27], the aim of the proposed methods is to investigate the performance of AdaFace on the challenging scenario presenting low quality occluded faces. The two proposed solutions (AdaFace4M and AdaFace12M) are trained using million-scale subsets of the cleaned version of the ultra-large-scale face benchmark consisting of 4Midentities/260M faces (Web-Face260M) [41]. AdaFace12M is trained on a larger subset of 12M of the curated WebFace42M.…”
Section: Biosmc-cdta Teammentioning
confidence: 99%
“…In the field of education, the wide application of IoT technology has brought about a sea change in teaching and learning. By connecting various devices, sensors and networks, IoT offers many new possibilities for education [ 2 ]. Teachers are utilizing IoT technology to better understand their progress and level of understanding of their students.…”
Section: Introductionmentioning
confidence: 99%