Fair Visual Recognition in Limited Data Regime using Self-Supervision
and Self-Distillation
- URL: http://arxiv.org/abs/2107.00067v1
- Date: Wed, 30 Jun 2021 19:22:46 GMT
- Title: Fair Visual Recognition in Limited Data Regime using Self-Supervision
and Self-Distillation
- Authors: Pratik Mazumder, Pravendra Singh, Vinay P. Namboodiri
- Abstract summary: We adapt self-supervision and self-distillation to reduce the impact of biases on the model.
We empirically show that our approach can significantly reduce the biases learned by the model.
Our approach significantly improves their performance and further reduces the model biases in the limited data regime.
- Score: 31.386413758098243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning models generally learn the biases present in the training data.
Researchers have proposed several approaches to mitigate such biases and make
the model fair. Bias mitigation techniques assume that a sufficiently large
number of training examples are present. However, we observe that if the
training data is limited, then the effectiveness of bias mitigation methods is
severely degraded. In this paper, we propose a novel approach to address this
problem. Specifically, we adapt self-supervision and self-distillation to
reduce the impact of biases on the model in this setting. Self-supervision and
self-distillation are not used for bias mitigation. However, through this work,
we demonstrate for the first time that these techniques are very effective in
bias mitigation. We empirically show that our approach can significantly reduce
the biases learned by the model. Further, we experimentally demonstrate that
our approach is complementary to other bias mitigation strategies. Our approach
significantly improves their performance and further reduces the model biases
in the limited data regime. Specifically, on the L-CIFAR-10S skewed dataset,
our approach significantly reduces the bias score of the baseline model by
78.22% and outperforms it in terms of accuracy by a significant absolute margin
of 8.89%. It also significantly reduces the bias score for the state-of-the-art
domain independent bias mitigation method by 59.26% and improves its
performance by a significant absolute margin of 7.08%.
Related papers
- How to be fair? A study of label and selection bias [3.018638214344819]
It is widely accepted that biased data leads to biased and potentially unfair models.
Several measures for bias in data and model predictions have been proposed, as well as bias mitigation techniques.
Despite the myriad of mitigation techniques developed in the past decade, it is still poorly understood under what circumstances which methods work.
arXiv Detail & Related papers (2024-03-21T10:43:55Z) - Hybrid Sample Synthesis-based Debiasing of Classifier in Limited Data
Setting [5.837881923712393]
This paper focuses on a more practical setting with no prior information about the bias.
In this setting, there are a large number of bias-aligned samples that cause the model to produce biased predictions.
If the training data is limited, the influence of the bias-aligned samples may become even stronger on the model predictions.
arXiv Detail & Related papers (2023-12-13T17:04:16Z) - Improving Bias Mitigation through Bias Experts in Natural Language
Understanding [10.363406065066538]
We propose a new debiasing framework that introduces binary classifiers between the auxiliary model and the main model.
Our proposed strategy improves the bias identification ability of the auxiliary model.
arXiv Detail & Related papers (2023-12-06T16:15:00Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - Language Models Get a Gender Makeover: Mitigating Gender Bias with
Few-Shot Data Interventions [50.67412723291881]
Societal biases present in pre-trained large language models are a critical issue.
We propose data intervention strategies as a powerful yet simple technique to reduce gender bias in pre-trained models.
arXiv Detail & Related papers (2023-06-07T16:50:03Z) - FITNESS: A Causal De-correlation Approach for Mitigating Bias in Machine
Learning Software [6.4073906779537095]
Biased datasets can lead to unfair and potentially harmful outcomes.
In this paper, we propose a bias mitigation approach via de-correlating the causal effects between sensitive features and the label.
Our key idea is that by de-correlating such effects from a causality perspective, the model would avoid making predictions based on sensitive features.
arXiv Detail & Related papers (2023-05-23T06:24:43Z) - Delving into Identify-Emphasize Paradigm for Combating Unknown Bias [52.76758938921129]
We propose an effective bias-conflicting scoring method (ECS) to boost the identification accuracy.
We also propose gradient alignment (GA) to balance the contributions of the mined bias-aligned and bias-conflicting samples.
Experiments are conducted on multiple datasets in various settings, demonstrating that the proposed solution can mitigate the impact of unknown biases.
arXiv Detail & Related papers (2023-02-22T14:50:24Z) - Feature-Level Debiased Natural Language Understanding [86.8751772146264]
Existing natural language understanding (NLU) models often rely on dataset biases to achieve high performance on specific datasets.
We propose debiasing contrastive learning (DCT) to mitigate biased latent features and neglect the dynamic nature of bias.
DCT outperforms state-of-the-art baselines on out-of-distribution datasets while maintaining in-distribution performance.
arXiv Detail & Related papers (2022-12-11T06:16:14Z) - Systematic Evaluation of Predictive Fairness [60.0947291284978]
Mitigating bias in training on biased datasets is an important open problem.
We examine the performance of various debiasing methods across multiple tasks.
We find that data conditions have a strong influence on relative model performance.
arXiv Detail & Related papers (2022-10-17T05:40:13Z) - Learning from others' mistakes: Avoiding dataset biases without modeling
them [111.17078939377313]
State-of-the-art natural language processing (NLP) models often learn to model dataset biases and surface form correlations instead of features that target the intended task.
Previous work has demonstrated effective methods to circumvent these issues when knowledge of the bias is available.
We show a method for training models that learn to ignore these problematic correlations.
arXiv Detail & Related papers (2020-12-02T16:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.