Semi-Siamese Training for Shallow Face Learning
- URL: http://arxiv.org/abs/2007.08398v1
- Date: Thu, 16 Jul 2020 15:20:04 GMT
- Title: Semi-Siamese Training for Shallow Face Learning
- Authors: Hang Du, Hailin Shi, Yuchi Liu, Jun Wang, Zhen Lei, Dan Zeng, Tao Mei
- Abstract summary: We introduce a novel training method named Semi-Siamese Training (SST)
A pair of Semi-Siamese networks constitute the forward propagation structure, and the training loss is computed with an updating gallery queue.
Our method is developed without extra-dependency, thus can be flexibly integrated with the existing loss functions and network architectures.
- Score: 78.7386209619276
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most existing public face datasets, such as MS-Celeb-1M and VGGFace2, provide
abundant information in both breadth (large number of IDs) and depth
(sufficient number of samples) for training. However, in many real-world
scenarios of face recognition, the training dataset is limited in depth, i.e.
only two face images are available for each ID. $\textit{We define this
situation as Shallow Face Learning, and find it problematic with existing
training methods.}$ Unlike deep face data, the shallow face data lacks
intra-class diversity. As such, it can lead to collapse of feature dimension
and consequently the learned network can easily suffer from degeneration and
over-fitting in the collapsed dimension. In this paper, we aim to address the
problem by introducing a novel training method named Semi-Siamese Training
(SST). A pair of Semi-Siamese networks constitute the forward propagation
structure, and the training loss is computed with an updating gallery queue,
conducting effective optimization on shallow training data. Our method is
developed without extra-dependency, thus can be flexibly integrated with the
existing loss functions and network architectures. Extensive experiments on
various benchmarks of face recognition show the proposed method significantly
improves the training, not only in shallow face learning, but also for
conventional deep face data.
Related papers
- Towards Metrical Reconstruction of Human Faces [20.782425305421505]
We argue for a supervised training scheme to learn the shape of a face.
We take advantage of a face recognition network pretrained on a large-scale 2D image dataset.
Our method outperforms the state-of-the-art reconstruction methods by a large margin.
arXiv Detail & Related papers (2022-04-13T18:57:33Z) - Multi-Agent Semi-Siamese Training for Long-tail and Shallow Face
Learning [54.13876727413492]
In many real-world scenarios of face recognition, the depth of training dataset is shallow, which means only two face images are available for each ID.
With the non-uniform increase of samples, such issue is converted to a more general case, a.k.a a long-tail face learning.
Based on the Semi-Siamese Training (SST), we introduce an advanced solution, named Multi-Agent Semi-Siamese Training (MASST)
MASST includes a probe network and multiple gallery agents, the former aims to encode the probe features, and the latter constitutes a stack of
arXiv Detail & Related papers (2021-05-10T04:57:32Z) - Recurrent Multi-view Alignment Network for Unsupervised Surface
Registration [79.72086524370819]
Learning non-rigid registration in an end-to-end manner is challenging due to the inherent high degrees of freedom and the lack of labeled training data.
We propose to represent the non-rigid transformation with a point-wise combination of several rigid transformations.
We also introduce a differentiable loss function that measures the 3D shape similarity on the projected multi-view 2D depth images.
arXiv Detail & Related papers (2020-11-24T14:22:42Z) - Towards Universal Representation Learning for Deep Face Recognition [106.21744671876704]
We propose a universal representation learning framework that can deal with larger variation unseen in the given training data without leveraging target domain knowledge.
Experiments show that our method achieves top performance on general face recognition datasets such as LFW and MegaFace.
arXiv Detail & Related papers (2020-02-26T23:29:57Z) - DotFAN: A Domain-transferred Face Augmentation Network for Pose and
Illumination Invariant Face Recognition [94.96686189033869]
We propose a 3D model-assisted domain-transferred face augmentation network (DotFAN)
DotFAN can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains.
Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity.
arXiv Detail & Related papers (2020-02-23T08:16:34Z) - Deep Multi-Facial Patches Aggregation Network For Facial Expression
Recognition [5.735035463793008]
We propose an approach for Facial Expressions Recognition (FER) based on a deep multi-facial patches aggregation network.
Deep features are learned from facial patches using deep sub-networks and aggregated within one deep architecture for expression classification.
arXiv Detail & Related papers (2020-02-20T17:57:06Z) - Joint Deep Learning of Facial Expression Synthesis and Recognition [97.19528464266824]
We propose a novel joint deep learning of facial expression synthesis and recognition method for effective FER.
The proposed method involves a two-stage learning procedure. Firstly, a facial expression synthesis generative adversarial network (FESGAN) is pre-trained to generate facial images with different facial expressions.
In order to alleviate the problem of data bias between the real images and the synthetic images, we propose an intra-class loss with a novel real data-guided back-propagation (RDBP) algorithm.
arXiv Detail & Related papers (2020-02-06T10:56:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.