EnD: Entangling and Disentangling deep representations for bias
correction
- URL: http://arxiv.org/abs/2103.02023v1
- Date: Tue, 2 Mar 2021 20:55:42 GMT
- Title: EnD: Entangling and Disentangling deep representations for bias
correction
- Authors: Enzo Tartaglione, Carlo Alberto Barbano, Marco Grangetto
- Abstract summary: We propose EnD, a regularization strategy whose aim is to prevent deep models from learning unwanted biases.
In particular, we insert an "information bottleneck" at a certain point of the deep neural network, where we disentangle the information about the bias.
Experiments show that EnD effectively improves the generalization on unbiased test sets.
- Score: 7.219077740523682
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Artificial neural networks perform state-of-the-art in an ever-growing number
of tasks, and nowadays they are used to solve an incredibly large variety of
tasks. There are problems, like the presence of biases in the training data,
which question the generalization capability of these models. In this work we
propose EnD, a regularization strategy whose aim is to prevent deep models from
learning unwanted biases. In particular, we insert an "information bottleneck"
at a certain point of the deep neural network, where we disentangle the
information about the bias, still letting the useful information for the
training task forward-propagating in the rest of the model. One big advantage
of EnD is that we do not require additional training complexity (like decoders
or extra layers in the model), since it is a regularizer directly applied on
the trained model. Our experiments show that EnD effectively improves the
generalization on unbiased test sets, and it can be effectively applied on
real-case scenarios, like removing hidden biases in the COVID-19 detection from
radiographic images.
Related papers
- Say My Name: a Model's Bias Discovery Framework [18.887645415907166]
We introduce Say My Name'' (SaMyNa), the first tool to identify biases within deep models semantically.
Unlike existing methods, our approach focuses on biases learned by the model.
Our method can disentangle task-related information and proposes itself as a tool to analyze biases.
arXiv Detail & Related papers (2024-08-18T18:50:59Z) - Model Debiasing by Learnable Data Augmentation [19.625915578646758]
This paper proposes a novel 2-stage learning pipeline featuring a data augmentation strategy able to regularize the training.
Experiments on synthetic and realistic biased datasets show state-of-the-art classification accuracy, outperforming competing methods.
arXiv Detail & Related papers (2024-08-09T09:19:59Z) - Combating Missing Modalities in Egocentric Videos at Test Time [92.38662956154256]
Real-world applications often face challenges with incomplete modalities due to privacy concerns, efficiency needs, or hardware issues.
We propose a novel approach to address this issue at test time without requiring retraining.
MiDl represents the first self-supervised, online solution for handling missing modalities exclusively at test time.
arXiv Detail & Related papers (2024-04-23T16:01:33Z) - Zero-shot Retrieval: Augmenting Pre-trained Models with Search Engines [83.65380507372483]
Large pre-trained models can dramatically reduce the amount of task-specific data required to solve a problem, but they often fail to capture domain-specific nuances out of the box.
This paper shows how to leverage recent advances in NLP and multi-modal learning to augment a pre-trained model with search engine retrieval.
arXiv Detail & Related papers (2023-11-29T05:33:28Z) - MENTOR: Human Perception-Guided Pretraining for Increased Generalization [5.596752018167751]
We introduce MENTOR (huMan pErceptioN-guided preTraining fOr increased geneRalization)
We train an autoencoder to learn human saliency maps given an input image, without class labels.
We remove the decoder part, add a classification layer on top of the encoder, and fine-tune this new model conventionally.
arXiv Detail & Related papers (2023-10-30T13:50:44Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - Relearning Forgotten Knowledge: on Forgetting, Overfit and Training-Free
Ensembles of DNNs [9.010643838773477]
We introduce a novel score for quantifying overfit, which monitors the forgetting rate of deep models on validation data.
We show that overfit can occur with and without a decrease in validation accuracy, and may be more common than previously appreciated.
We use our observations to construct a new ensemble method, based solely on the training history of a single network, which provides significant improvement without any additional cost in training time.
arXiv Detail & Related papers (2023-10-17T09:22:22Z) - Toward More Generalized Malicious URL Detection Models [4.151658495779136]
This paper reveals a data bias issue that can severely affect the performance while conducting a machine learning model for malicious URL detection.
We describe how such bias can be identified using interpretable machine learning techniques, and argue that such biases naturally exist in the real world security data for training a classification model.
We propose a debiased training strategy that can be applied to most deep-learning based models to alleviate the negative effects from the biased features.
arXiv Detail & Related papers (2022-02-21T07:46:03Z) - General Greedy De-bias Learning [163.65789778416172]
We propose a General Greedy De-bias learning framework (GGD), which greedily trains the biased models and the base model like gradient descent in functional space.
GGD can learn a more robust base model under the settings of both task-specific biased models with prior knowledge and self-ensemble biased model without prior knowledge.
arXiv Detail & Related papers (2021-12-20T14:47:32Z) - Reasoning-Modulated Representations [85.08205744191078]
We study a common setting where our task is not purely opaque.
Our approach paves the way for a new class of data-efficient representation learning.
arXiv Detail & Related papers (2021-07-19T13:57:13Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.