Simon Says: Evaluating and Mitigating Bias in Pruned Neural Networks
with Knowledge Distillation
- URL: http://arxiv.org/abs/2106.07849v1
- Date: Tue, 15 Jun 2021 02:59:32 GMT
- Title: Simon Says: Evaluating and Mitigating Bias in Pruned Neural Networks
with Knowledge Distillation
- Authors: Cody Blakeney, Nathaniel Huish, Yan Yan, Ziliang Zong
- Abstract summary: A clear gap exists in the current literature on evaluating and mitigating bias in pruned neural networks.
We propose two simple yet effective metrics, Combined Error Variance (CEV) and Symmetric Distance Error (SDE) to quantitatively evaluate the induced bias prevention quality.
Second, we demonstrate that knowledge distillation can mitigate induced bias in pruned neural networks, even with unbalanced datasets.
Third, we reveal that model similarity has strong correlations with pruning induced bias, which provides a powerful method to explain why bias occurs in pruned neural networks.
- Score: 8.238238958749134
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In recent years the ubiquitous deployment of AI has posed great concerns in
regards to algorithmic bias, discrimination, and fairness. Compared to
traditional forms of bias or discrimination caused by humans, algorithmic bias
generated by AI is more abstract and unintuitive therefore more difficult to
explain and mitigate. A clear gap exists in the current literature on
evaluating and mitigating bias in pruned neural networks. In this work, we
strive to tackle the challenging issues of evaluating, mitigating, and
explaining induced bias in pruned neural networks. Our paper makes three
contributions. First, we propose two simple yet effective metrics, Combined
Error Variance (CEV) and Symmetric Distance Error (SDE), to quantitatively
evaluate the induced bias prevention quality of pruned models. Second, we
demonstrate that knowledge distillation can mitigate induced bias in pruned
neural networks, even with unbalanced datasets. Third, we reveal that model
similarity has strong correlations with pruning induced bias, which provides a
powerful method to explain why bias occurs in pruned neural networks. Our code
is available at https://github.com/codestar12/pruning-distilation-bias
Related papers
- Debiasify: Self-Distillation for Unsupervised Bias Mitigation [19.813054813868476]
Simplicity bias poses a significant challenge in neural networks, often leading models to favor simpler solutions and inadvertently learn decision rules influenced by spurious correlations.
We introduce Debiasify, a novel self-distillation approach that requires no prior knowledge about the nature of biases.
Our method leverages a new distillation loss to transfer knowledge within the network, from deeper layers containing complex, highly-predictive features to shallower layers with simpler, attribute-conditioned features in an unsupervised manner.
arXiv Detail & Related papers (2024-11-01T16:25:05Z) - NeuFair: Neural Network Fairness Repair with Dropout [19.49034966552718]
This paper investigates neuron dropout as a post-processing bias mitigation for deep neural networks (DNNs)
We show that our design of randomized algorithms is effective and efficient in improving fairness (up to 69%) with minimal or no model performance degradation.
arXiv Detail & Related papers (2024-07-05T05:45:34Z) - Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures [93.17009514112702]
Pruning, setting a significant subset of the parameters of a neural network to zero, is one of the most popular methods of model compression.
Despite existing evidence for this phenomenon, the relationship between neural network pruning and induced bias is not well-understood.
arXiv Detail & Related papers (2023-04-25T07:42:06Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Unsupervised Learning of Unbiased Visual Representations [10.871587311621974]
Deep neural networks are known for their inability to learn robust representations when biases exist in the dataset.
We propose a fully unsupervised debiasing framework, consisting of three steps.
We employ state-of-the-art supervised debiasing techniques to obtain an unbiased model.
arXiv Detail & Related papers (2022-04-26T10:51:50Z) - Vulnerability Under Adversarial Machine Learning: Bias or Variance? [77.30759061082085]
We investigate the effect of adversarial machine learning on the bias and variance of a trained deep neural network.
Our analysis sheds light on why the deep neural networks have poor performance under adversarial perturbation.
We introduce a new adversarial machine learning algorithm with lower computational complexity than well-known adversarial machine learning strategies.
arXiv Detail & Related papers (2020-08-01T00:58:54Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z) - Fairness Through Robustness: Investigating Robustness Disparity in Deep
Learning [61.93730166203915]
We argue that traditional notions of fairness are not sufficient when the model is vulnerable to adversarial attacks.
We show that measuring robustness bias is a challenging task for DNNs and propose two methods to measure this form of bias.
arXiv Detail & Related papers (2020-06-17T22:22:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.