Omni-supervised Facial Expression Recognition via Distilled Data
- URL: http://arxiv.org/abs/2005.08551v5
- Date: Thu, 9 Dec 2021 04:07:04 GMT
- Title: Omni-supervised Facial Expression Recognition via Distilled Data
- Authors: Ping Liu, Yunchao Wei, Zibo Meng, Weihong Deng, Joey Tianyi Zhou, Yi
Yang
- Abstract summary: We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
- Score: 120.11782405714234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Facial expression plays an important role in understanding human emotions.
Most recently, deep learning based methods have shown promising for facial
expression recognition. However, the performance of the current
state-of-the-art facial expression recognition (FER) approaches is directly
related to the labeled data for training. To solve this issue, prior works
employ the pretrain-and-finetune strategy, i.e., utilize a large amount of
unlabeled data to pretrain the network and then finetune it by the labeled
data. As the labeled data is in a small amount, the final network performance
is still restricted. From a different perspective, we propose to perform
omni-supervised learning to directly exploit reliable samples in a large amount
of unlabeled data for network training. Particularly, a new dataset is firstly
constructed using a primitive model trained on a small number of labeled
samples to select samples with high confidence scores from a face dataset,
i.e., MS-Celeb-1M, based on feature-wise similarity. We experimentally verify
that the new dataset created in such an omni-supervised manner can
significantly improve the generalization ability of the learned FER model.
However, as the number of training samples grows, computational cost and
training time increase dramatically. To tackle this, we propose to apply a
dataset distillation strategy to compress the created dataset into several
informative class-wise images, significantly improving the training efficiency.
We have conducted extensive experiments on widely used benchmarks, where
consistent performance gains can be achieved under various settings using the
proposed framework. More importantly, the distilled dataset has shown its
capabilities of boosting the performance of FER with negligible additional
computational costs.
Related papers
- In-Domain Self-Supervised Learning Improves Remote Sensing Image Scene
Classification [5.323049242720532]
Self-supervised learning has emerged as a promising approach for remote sensing image classification.
We present a study of different self-supervised pre-training strategies and evaluate their effect across 14 downstream datasets.
arXiv Detail & Related papers (2023-07-04T10:57:52Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Self-Supervised Pre-Training for Transformer-Based Person
Re-Identification [54.55281692768765]
Transformer-based supervised pre-training achieves great performance in person re-identification (ReID)
Due to the domain gap between ImageNet and ReID datasets, it usually needs a larger pre-training dataset to boost the performance.
This work aims to mitigate the gap between the pre-training and ReID datasets from the perspective of data and model structure.
arXiv Detail & Related papers (2021-11-23T18:59:08Z) - SelfHAR: Improving Human Activity Recognition through Self-training with
Unlabeled Data [9.270269467155547]
SelfHAR is a semi-supervised model that learns to leverage unlabeled datasets to complement small labeled datasets.
Our approach combines teacher-student self-training, which distills the knowledge of unlabeled and labeled datasets.
SelfHAR is data-efficient, reaching similar performance using up to 10 times less labeled data compared to supervised approaches.
arXiv Detail & Related papers (2021-02-11T15:40:35Z) - Data Impressions: Mining Deep Models to Extract Samples for Data-free
Applications [26.48630545028405]
"Data Impressions" act as proxy to the training data and can be used to realize a variety of tasks.
We show the applicability of data impressions in solving several computer vision tasks.
arXiv Detail & Related papers (2021-01-15T11:37:29Z) - Noisy Student Training using Body Language Dataset Improves Facial
Expression Recognition [10.529781894367877]
In this paper, we use a self-training method that utilizes a combination of a labelled dataset and an unlabelled dataset.
Experimental analysis shows that training a noisy student network iteratively helps in achieving significantly better results.
Our results show that the proposed method achieves state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2020-08-06T13:45:52Z) - Learning to Count in the Crowd from Limited Labeled Data [109.2954525909007]
We focus on reducing the annotation efforts by learning to count in the crowd from limited number of labeled samples.
Specifically, we propose a Gaussian Process-based iterative learning mechanism that involves estimation of pseudo-ground truth for the unlabeled data.
arXiv Detail & Related papers (2020-07-07T04:17:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.