WebFace260M: A Benchmark for Million-Scale Deep Face Recognition
- URL: http://arxiv.org/abs/2204.10149v1
- Date: Thu, 21 Apr 2022 14:56:53 GMT
- Title: WebFace260M: A Benchmark for Million-Scale Deep Face Recognition
- Authors: Zheng Zhu, Guan Huang, Jiankang Deng, Yun Ye, Junjie Huang, Xinze
Chen, Jiagang Zhu, Tian Yang, Dalong Du, Jiwen Lu, Jie Zhou
- Abstract summary: We contribute a new million-scale recognition benchmark, containing uncurated 4M identities/260M faces (WebFace260M) and cleaned 2M identities/42M faces (WebFace42M)
A distributed framework is developed to train face recognition models efficiently without tampering with the performance.
The proposed benchmark shows enormous potential on standard, masked and unbiased face recognition scenarios.
- Score: 89.39080252029386
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Face benchmarks empower the research community to train and evaluate
high-performance face recognition systems. In this paper, we contribute a new
million-scale recognition benchmark, containing uncurated 4M identities/260M
faces (WebFace260M) and cleaned 2M identities/42M faces (WebFace42M) training
data, as well as an elaborately designed time-constrained evaluation protocol.
Firstly, we collect 4M name lists and download 260M faces from the Internet.
Then, a Cleaning Automatically utilizing Self-Training (CAST) pipeline is
devised to purify the tremendous WebFace260M, which is efficient and scalable.
To the best of our knowledge, the cleaned WebFace42M is the largest public face
recognition training set and we expect to close the data gap between academia
and industry. Referring to practical deployments, Face Recognition Under
Inference Time conStraint (FRUITS) protocol and a new test set with rich
attributes are constructed. Besides, we gather a large-scale masked face
sub-set for biometrics assessment under COVID-19. For a comprehensive
evaluation of face matchers, three recognition tasks are performed under
standard, masked and unbiased settings, respectively. Equipped with this
benchmark, we delve into million-scale face recognition problems. A distributed
framework is developed to train face recognition models efficiently without
tampering with the performance. Enabled by WebFace42M, we reduce 40% failure
rate on the challenging IJB-C set and rank 3rd among 430 entries on NIST-FRVT.
Even 10% data (WebFace4M) shows superior performance compared with the public
training sets. Furthermore, comprehensive baselines are established under the
FRUITS-100/500/1000 milliseconds protocols. The proposed benchmark shows
enormous potential on standard, masked and unbiased face recognition scenarios.
Our WebFace260M website is https://www.face-benchmark.org.
Related papers
- How to Boost Face Recognition with StyleGAN? [13.067766076889995]
State-of-the-art face recognition systems require vast amounts of labeled training data.
Self-supervised revolution in the industry motivates research on the adaptation of related techniques to facial recognition.
We show that a simple approach based on fine-tuning pSp encoder for StyleGAN allows us to improve upon the state-of-the-art facial recognition.
arXiv Detail & Related papers (2022-10-18T18:41:56Z) - Blind Face Restoration: Benchmark Datasets and a Baseline Model [63.053331687284064]
Blind Face Restoration (BFR) aims to construct a high-quality (HQ) face image from its corresponding low-quality (LQ) input.
We first synthesize two blind face restoration benchmark datasets called EDFace-Celeb-1M (BFR128) and EDFace-Celeb-150K (BFR512)
State-of-the-art methods are benchmarked on them under five settings including blur, noise, low resolution, JPEG compression artifacts, and the combination of them (full degradation)
arXiv Detail & Related papers (2022-06-08T06:34:24Z) - FaceMAE: Privacy-Preserving Face Recognition via Masked Autoencoders [81.21440457805932]
We propose a novel framework FaceMAE, where the face privacy and recognition performance are considered simultaneously.
randomly masked face images are used to train the reconstruction module in FaceMAE.
We also perform sufficient privacy-preserving face recognition on several public face datasets.
arXiv Detail & Related papers (2022-05-23T07:19:42Z) - Face-NMS: A Core-set Selection Approach for Efficient Face Recognition [14.863570260332747]
Face recognition in the wild has achieved remarkable success.
One key engine is the increasing size of training data.
Massive number of faces raise the constraints in training time, computing resources, and memory cost.
arXiv Detail & Related papers (2021-09-10T07:07:04Z) - Masked Face Recognition Challenge: The InsightFace Track Report [79.77020394722788]
During the COVID-19 coronavirus epidemic, almost everyone wears a facial mask, which poses a huge challenge to deep face recognition.
In this workshop, we focus on bench-marking deep face recognition methods under the existence of facial masks.
arXiv Detail & Related papers (2021-08-18T15:14:44Z) - Masked Face Recognition Challenge: The WebFace260M Track Report [81.57455766506197]
Face Bio-metrics under COVID Workshop and Masked Face Recognition Challenge in ICCV 2021.
WebFace260M Track aims to push the frontiers of practical MFR.
In the first phase of WebFace260M Track, 69 teams (total 833 solutions) participate in the challenge.
There are second phase of the challenge till October 1, 2021 and on-going leaderboard.
arXiv Detail & Related papers (2021-08-16T15:51:51Z) - WebFace260M: A Benchmark Unveiling the Power of Million-Scale Deep Face
Recognition [79.65728162193584]
We contribute a new million-scale face benchmark containing noisy 4M identities/260M faces (WebFace260M) and cleaned 2M identities/42M faces (WebFace42M)
We reduce relative 40% failure rate on the challenging IJB-C set, and ranks the 3rd among 430 entries on NIST-FRVT.
Even 10% data (WebFace4M) shows superior performance compared with public training set.
arXiv Detail & Related papers (2021-03-06T11:12:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.