Consensual Collaborative Training And Knowledge Distillation Based
Facial Expression Recognition Under Noisy Annotations
- URL: http://arxiv.org/abs/2107.04746v1
- Date: Sat, 10 Jul 2021 03:37:06 GMT
- Title: Consensual Collaborative Training And Knowledge Distillation Based
Facial Expression Recognition Under Noisy Annotations
- Authors: Darshan Gera, S. Balasubramanian
- Abstract summary: This work proposes an effective training strategy in the presence of noisy labels, called as Consensual Collaborative Training (CCT) framework.
CCT co-trains three networks jointly using a convex combination of supervision loss and consistency loss.
State-of-the-art performance is reported on the benchmark FER datasets RAFDB (90.84%), FERPlus (89.99%) and AffectNet (66%)
- Score: 2.538209532048867
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Presence of noise in the labels of large scale facial expression datasets has
been a key challenge towards Facial Expression Recognition (FER) in the wild.
During early learning stage, deep networks fit on clean data. Then, eventually,
they start overfitting on noisy labels due to their memorization ability, which
limits FER performance. This work proposes an effective training strategy in
the presence of noisy labels, called as Consensual Collaborative Training (CCT)
framework. CCT co-trains three networks jointly using a convex combination of
supervision loss and consistency loss, without making any assumption about the
noise distribution. A dynamic transition mechanism is used to move from
supervision loss in early learning to consistency loss for consensus of
predictions among networks in the later stage. Inference is done using a single
network based on a simple knowledge distillation scheme. Effectiveness of the
proposed framework is demonstrated on synthetic as well as real noisy FER
datasets. In addition, a large test subset of around 5K images is annotated
from the FEC dataset using crowd wisdom of 16 different annotators and reliable
labels are inferred. CCT is also validated on it. State-of-the-art performance
is reported on the benchmark FER datasets RAFDB (90.84%) FERPlus (89.99%) and
AffectNet (66%). Our codes are available at https://github.com/1980x/CCT.
Related papers
- SSP-RACL: Classification of Noisy Fundus Images with Self-Supervised Pretraining and Robust Adaptive Credal Loss [3.8739860035485143]
Fundus image classification is crucial in the computer aided diagnosis tasks, but label noise significantly impairs the performance of deep neural networks.
We propose a robust framework, Self-Supervised Pre-training with Robust Adaptive Credal Loss (SSP-RACL), for handling label noise in fundus image datasets.
arXiv Detail & Related papers (2024-09-25T02:41:58Z) - ERASE: Error-Resilient Representation Learning on Graphs for Label Noise
Tolerance [53.73316938815873]
We propose a method called ERASE (Error-Resilient representation learning on graphs for lAbel noiSe tolerancE) to learn representations with error tolerance.
ERASE combines prototype pseudo-labels with propagated denoised labels and updates representations with error resilience.
Our method can outperform multiple baselines with clear margins in broad noise levels and enjoy great scalability.
arXiv Detail & Related papers (2023-12-13T17:59:07Z) - Dynamic Adaptive Threshold based Learning for Noisy Annotations Robust
Facial Expression Recognition [3.823356975862006]
We propose a dynamic FER learning framework (DNFER) to handle noisy annotations.
Specifically, DNFER is based on supervised training using selected clean samples and unsupervised consistent training using all the samples.
We demonstrate the robustness of DNFER on both synthetic as well as on real noisy annotated FER datasets like RAFDB, FERPlus, SFEW and AffectNet.
arXiv Detail & Related papers (2022-08-22T12:02:41Z) - Boosting Facial Expression Recognition by A Semi-Supervised Progressive
Teacher [54.50747989860957]
We propose a semi-supervised learning algorithm named Progressive Teacher (PT) to utilize reliable FER datasets as well as large-scale unlabeled expression images for effective training.
Experiments on widely-used databases RAF-DB and FERPlus validate the effectiveness of our method, which achieves state-of-the-art performance with accuracy of 89.57% on RAF-DB.
arXiv Detail & Related papers (2022-05-28T07:47:53Z) - Synergistic Network Learning and Label Correction for Noise-robust Image
Classification [28.27739181560233]
Deep Neural Networks (DNNs) tend to overfit training label noise, resulting in poorer model performance in practice.
We propose a robust label correction framework combining the ideas of small loss selection and noise correction.
We demonstrate our method on both synthetic and real-world datasets with different noise types and rates.
arXiv Detail & Related papers (2022-02-27T23:06:31Z) - Consistency Regularization Can Improve Robustness to Label Noise [4.340338299803562]
This paper empirically studies the relevance of consistency regularization for training-time robustness to noisy labels.
We show that a simple loss function that encourages consistency improves the robustness of the models to label noise.
arXiv Detail & Related papers (2021-10-04T08:15:08Z) - Affect Expression Behaviour Analysis in the Wild using Consensual
Collaborative Training [2.538209532048867]
This report presents Consensual Collaborative Training (CCT) framework used in our submission to the Affective Behaviour Analysis in-the-wild (ABAW) 2021 competition.
CCT co-trains three networks jointly using a convex combination of supervision loss and consistency loss.
Co-training reduces overall error, and consistency loss prevents overfitting to noisy samples.
arXiv Detail & Related papers (2021-07-08T04:28:21Z) - SCARF: Self-Supervised Contrastive Learning using Random Feature
Corruption [72.35532598131176]
We propose SCARF, a technique for contrastive learning, where views are formed by corrupting a random subset of features.
We show that SCARF complements existing strategies and outperforms alternatives like autoencoders.
arXiv Detail & Related papers (2021-06-29T08:08:33Z) - Coresets for Robust Training of Neural Networks against Noisy Labels [78.03027938765746]
We propose a novel approach with strong theoretical guarantees for robust training of deep networks trained with noisy labels.
We select weighted subsets (coresets) of clean data points that provide an approximately low-rank Jacobian matrix.
Our experiments corroborate our theory and demonstrate that deep networks trained on our subsets achieve a significantly superior performance compared to state-of-the art.
arXiv Detail & Related papers (2020-11-15T04:58:11Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - Suppressing Uncertainties for Large-Scale Facial Expression Recognition [81.51495681011404]
This paper proposes a simple yet efficient Self-Cure Network (SCN) which suppresses the uncertainties efficiently and prevents deep networks from over-fitting uncertain facial images.
Results on public benchmarks demonstrate that our SCN outperforms current state-of-the-art methods with textbf88.14% on RAF-DB, textbf60.23% on AffectNet, and textbf89.35% on FERPlus.
arXiv Detail & Related papers (2020-02-24T17:24:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.