Improving Face Recognition from Hard Samples via Distribution
Distillation Loss
- URL: http://arxiv.org/abs/2002.03662v3
- Date: Sat, 18 Jul 2020 15:00:35 GMT
- Title: Improving Face Recognition from Hard Samples via Distribution
Distillation Loss
- Authors: Yuge Huang, Pengcheng Shen, Ying Tai, Shaoxin Li, Xiaoming Liu, Jilin
Li, Feiyue Huang, Rongrong Ji
- Abstract summary: Large facial variations are the main challenge in face recognition.
We propose a novel Distribution Distillation Loss to narrow the performance gap between easy and hard samples.
We have conducted extensive experiments on both generic large-scale face benchmarks and benchmarks with diverse variations on race, resolution and pose.
- Score: 131.61036519863856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large facial variations are the main challenge in face recognition. To this
end, previous variation-specific methods make full use of task-related prior to
design special network losses, which are typically not general among different
tasks and scenarios. In contrast, the existing generic methods focus on
improving the feature discriminability to minimize the intra-class distance
while maximizing the interclass distance, which perform well on easy samples
but fail on hard samples. To improve the performance on those hard samples for
general tasks, we propose a novel Distribution Distillation Loss to narrow the
performance gap between easy and hard samples, which is a simple, effective and
generic for various types of facial variations. Specifically, we first adopt
state-of-the-art classifiers such as ArcFace to construct two similarity
distributions: teacher distribution from easy samples and student distribution
from hard samples. Then, we propose a novel distribution-driven loss to
constrain the student distribution to approximate the teacher distribution,
which thus leads to smaller overlap between the positive and negative pairs in
the student distribution. We have conducted extensive experiments on both
generic large-scale face benchmarks and benchmarks with diverse variations on
race, resolution and pose. The quantitative results demonstrate the superiority
of our method over strong baselines, e.g., Arcface and Cosface.
Related papers
- DiscrimLoss: A Universal Loss for Hard Samples and Incorrect Samples
Discrimination [28.599571524763785]
Given data with label noise (i.e., incorrect data), deep neural networks would gradually memorize the label noise and impair model performance.
To relieve this issue, curriculum learning is proposed to improve model performance and generalization by ordering training samples in a meaningful sequence.
arXiv Detail & Related papers (2022-08-21T13:38:55Z) - Few-shot Forgery Detection via Guided Adversarial Interpolation [56.59499187594308]
Existing forgery detection methods suffer from significant performance drops when applied to unseen novel forgery approaches.
We propose Guided Adversarial Interpolation (GAI) to overcome the few-shot forgery detection problem.
Our method is validated to be robust to choices of majority and minority forgery approaches.
arXiv Detail & Related papers (2022-04-12T16:05:10Z) - OTFace: Hard Samples Guided Optimal Transport Loss for Deep Face
Representation [31.220594076407444]
Face representation in the wild is extremely hard due to the large scale face variations.
This paper proposes the hard samples guided optimal transport (OT) loss for deep face representation, OTFace.
arXiv Detail & Related papers (2022-03-28T02:57:04Z) - Jo-SRC: A Contrastive Approach for Combating Noisy Labels [58.867237220886885]
We propose a noise-robust approach named Jo-SRC (Joint Sample Selection and Model Regularization based on Consistency)
Specifically, we train the network in a contrastive learning manner. Predictions from two different views of each sample are used to estimate its "likelihood" of being clean or out-of-distribution.
arXiv Detail & Related papers (2021-03-24T07:26:07Z) - CurricularFace: Adaptive Curriculum Learning Loss for Deep Face
Recognition [79.92240030758575]
We propose a novel Adaptive Curriculum Learning loss (CurricularFace) that embeds the idea of curriculum learning into the loss function.
Our CurricularFace adaptively adjusts the relative importance of easy and hard samples during different training stages.
arXiv Detail & Related papers (2020-04-01T08:43:10Z) - One-Shot Domain Adaptation For Face Generation [34.882820002799626]
We propose a framework capable of generating face images that fall into the same distribution as that of a given one-shot example.
We develop an iterative optimization scheme that rapidly adapts the weights of the model to shift the output's high-level distribution to the target's.
To generate images of the same distribution, we introduce a style-mixing technique that transfers the low-level statistics from the target to faces randomly generated with the model.
arXiv Detail & Related papers (2020-03-28T18:50:13Z) - Towards Universal Representation Learning for Deep Face Recognition [106.21744671876704]
We propose a universal representation learning framework that can deal with larger variation unseen in the given training data without leveraging target domain knowledge.
Experiments show that our method achieves top performance on general face recognition datasets such as LFW and MegaFace.
arXiv Detail & Related papers (2020-02-26T23:29:57Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.