Active Learning with Contrastive Pre-training for Facial Expression
Recognition
- URL: http://arxiv.org/abs/2307.02744v1
- Date: Thu, 6 Jul 2023 03:08:03 GMT
- Title: Active Learning with Contrastive Pre-training for Facial Expression
Recognition
- Authors: Shuvendu Roy, Ali Etemad
- Abstract summary: We study 8 recent active learning methods on three public FER datasets.
Our findings show that existing active learning methods do not perform well in the context of FER.
We propose contrastive self-supervised pre-training, which first learns the underlying representations based on the entire unlabelled dataset.
- Score: 19.442685015494316
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Deep learning has played a significant role in the success of facial
expression recognition (FER), thanks to large models and vast amounts of
labelled data. However, obtaining labelled data requires a tremendous amount of
human effort, time, and financial resources. Even though some prior works have
focused on reducing the need for large amounts of labelled data using different
unsupervised methods, another promising approach called active learning is
barely explored in the context of FER. This approach involves selecting and
labelling the most representative samples from an unlabelled set to make the
best use of a limited 'labelling budget'. In this paper, we implement and study
8 recent active learning methods on three public FER datasets, FER13, RAF-DB,
and KDEF. Our findings show that existing active learning methods do not
perform well in the context of FER, likely suffering from a phenomenon called
'Cold Start', which occurs when the initial set of labelled samples is not well
representative of the entire dataset. To address this issue, we propose
contrastive self-supervised pre-training, which first learns the underlying
representations based on the entire unlabelled dataset. We then follow this
with the active learning methods and observe that our 2-step approach shows up
to 9.2% improvement over random sampling and up to 6.7% improvement over the
best existing active learning baseline without the pre-training. We will make
the code for this study public upon publication at:
github.com/ShuvenduRoy/ActiveFER.
Related papers
- Self-Training for Sample-Efficient Active Learning for Text Classification with Pre-Trained Language Models [3.546617486894182]
We introduce HAST, a new and effective self-training strategy, which is evaluated on four text classification benchmarks.
Results show that it outperforms the reproduced self-training approaches and reaches classification results comparable to previous experiments for three out of four datasets.
arXiv Detail & Related papers (2024-06-13T15:06:11Z) - Incremental Self-training for Semi-supervised Learning [56.57057576885672]
IST is simple yet effective and fits existing self-training-based semi-supervised learning methods.
We verify the proposed IST on five datasets and two types of backbone, effectively improving the recognition accuracy and learning speed.
arXiv Detail & Related papers (2024-04-14T05:02:00Z) - XAL: EXplainable Active Learning Makes Classifiers Better Low-resource Learners [71.8257151788923]
We propose a novel Explainable Active Learning framework (XAL) for low-resource text classification.
XAL encourages classifiers to justify their inferences and delve into unlabeled data for which they cannot provide reasonable explanations.
Experiments on six datasets show that XAL achieves consistent improvement over 9 strong baselines.
arXiv Detail & Related papers (2023-10-09T08:07:04Z) - Exploring the Boundaries of Semi-Supervised Facial Expression
Recognition: Learning from In-Distribution, Out-of-Distribution, and
Unconstrained Data [19.442685015494316]
We present a study on 11 of the most recent semi-supervised methods, in the context of facial expression recognition (FER)
Our investigation covers semi-supervised learning from in-distribution, out-of-distribution, unconstrained, and very small unlabelled data.
Our results demonstrate that FixMatch consistently achieves better performance on in-distribution unlabelled data, while ReMixMatch stands out among all methods for out-of-distribution, unconstrained, and scarce unlabelled data scenarios.
arXiv Detail & Related papers (2023-06-02T01:40:08Z) - Responsible Active Learning via Human-in-the-loop Peer Study [88.01358655203441]
We propose a responsible active learning method, namely Peer Study Learning (PSL), to simultaneously preserve data privacy and improve model stability.
We first introduce a human-in-the-loop teacher-student architecture to isolate unlabelled data from the task learner (teacher) on the cloud-side.
During training, the task learner instructs the light-weight active learner which then provides feedback on the active sampling criterion.
arXiv Detail & Related papers (2022-11-24T13:18:27Z) - SURF: Semi-supervised Reward Learning with Data Augmentation for
Feedback-efficient Preference-based Reinforcement Learning [168.89470249446023]
We present SURF, a semi-supervised reward learning framework that utilizes a large amount of unlabeled samples with data augmentation.
In order to leverage unlabeled samples for reward learning, we infer pseudo-labels of the unlabeled samples based on the confidence of the preference predictor.
Our experiments demonstrate that our approach significantly improves the feedback-efficiency of the preference-based method on a variety of locomotion and robotic manipulation tasks.
arXiv Detail & Related papers (2022-03-18T16:50:38Z) - Towards Reducing Labeling Cost in Deep Object Detection [61.010693873330446]
We propose a unified framework for active learning, that considers both the uncertainty and the robustness of the detector.
Our method is able to pseudo-label the very confident predictions, suppressing a potential distribution drift.
arXiv Detail & Related papers (2021-06-22T16:53:09Z) - Ask-n-Learn: Active Learning via Reliable Gradient Representations for
Image Classification [29.43017692274488]
Deep predictive models rely on human supervision in the form of labeled training data.
We propose Ask-n-Learn, an active learning approach based on gradient embeddings obtained using the pesudo-labels estimated in each of the algorithm.
arXiv Detail & Related papers (2020-09-30T05:19:56Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.