Killing Two Birds with One Stone:Efficient and Robust Training of Face
Recognition CNNs by Partial FC
- URL: http://arxiv.org/abs/2203.15565v1
- Date: Mon, 28 Mar 2022 14:33:21 GMT
- Title: Killing Two Birds with One Stone:Efficient and Robust Training of Face
Recognition CNNs by Partial FC
- Authors: Xiang An and Jiankang Deng and Jia Guo and Ziyong Feng and Xuhan Zhu
and Jing Yang and Tongliang Liu
- Abstract summary: We propose a sparsely updating variant of the Fully Connected (FC) layer, named Partial FC (PFC)
In each iteration, positive class centers and a random subset of negative class centers are selected to compute the margin-based softmax loss.
The computing requirement, the probability of inter-class conflict, and the frequency of passive update on tail class centers, are dramatically reduced.
- Score: 66.71660672526349
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning discriminative deep feature embeddings by using million-scale
in-the-wild datasets and margin-based softmax loss is the current
state-of-the-art approach for face recognition. However, the memory and
computing cost of the Fully Connected (FC) layer linearly scales up to the
number of identities in the training set. Besides, the large-scale training
data inevitably suffers from inter-class conflict and long-tailed distribution.
In this paper, we propose a sparsely updating variant of the FC layer, named
Partial FC (PFC). In each iteration, positive class centers and a random subset
of negative class centers are selected to compute the margin-based softmax
loss. All class centers are still maintained throughout the whole training
process, but only a subset is selected and updated in each iteration.
Therefore, the computing requirement, the probability of inter-class conflict,
and the frequency of passive update on tail class centers, are dramatically
reduced. Extensive experiments across different training data and backbones
(e.g. CNN and ViT) confirm the effectiveness, robustness and efficiency of the
proposed PFC. The source code is available at
\https://github.com/deepinsight/insightface/tree/master/recognition.
Related papers
- Federated Learning under Partially Class-Disjoint Data via Manifold Reshaping [64.58402571292723]
We propose a manifold reshaping approach called FedMR to calibrate the feature space of local training.
We conduct extensive experiments on a range of datasets to demonstrate that our FedMR achieves much higher accuracy and better communication efficiency.
arXiv Detail & Related papers (2024-05-29T10:56:13Z) - Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - Class Instance Balanced Learning for Long-Tailed Classification [0.0]
Long-tailed image classification task deals with large imbalances in the class frequencies of the training data.
Previous approaches have shown that combining cross-entropy and contrastive learning can improve performance on the long-tailed task.
We propose a novel class instance balanced loss (CIBL), which reweights the relative contributions of a cross-entropy and a contrastive loss as a function of the frequency of class instances in the training batch.
arXiv Detail & Related papers (2023-07-11T15:09:10Z) - Addressing Class Variable Imbalance in Federated Semi-supervised
Learning [10.542178602467885]
We propose Federated Semi-supervised Learning for Class Variable Imbalance (FCVI) to solve class variable imbalance.
FCVI is used to mitigate the data imbalance due to changes of the number of classes.
Our scheme is proved to be significantly better than baseline methods, while maintaining client privacy.
arXiv Detail & Related papers (2023-03-21T12:50:17Z) - R-DFCIL: Relation-Guided Representation Learning for Data-Free Class
Incremental Learning [64.7996065569457]
Class-Incremental Learning (CIL) struggles with catastrophic forgetting when learning new knowledge.
Recent DFCIL works introduce techniques such as model inversion to synthesize data for previous classes, they fail to overcome forgetting due to the severe domain gap between the synthetic and real data.
This paper proposes relation-guided representation learning (RRL) for DFCIL, dubbed R-DFCIL.
arXiv Detail & Related papers (2022-03-24T14:54:15Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Acceleration of Federated Learning with Alleviated Forgetting in Local
Training [61.231021417674235]
Federated learning (FL) enables distributed optimization of machine learning models while protecting privacy.
We propose FedReg, an algorithm to accelerate FL with alleviated knowledge forgetting in the local training stage.
Our experiments demonstrate that FedReg not only significantly improves the convergence rate of FL, especially when the neural network architecture is deep.
arXiv Detail & Related papers (2022-03-05T02:31:32Z) - You Only Need End-to-End Training for Long-Tailed Recognition [8.789819609485225]
Cross-entropy loss tends to produce highly correlated features on imbalanced data.
We propose two novel modules, Block-based Relatively Balanced Batch Sampler (B3RS) and Batch Embedded Training (BET)
Experimental results on the long-tailed classification benchmarks, CIFAR-LT and ImageNet-LT, demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2021-12-11T11:44:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.