Discover and Cure: Concept-aware Mitigation of Spurious Correlation
- URL: http://arxiv.org/abs/2305.00650v2
- Date: Mon, 5 Jun 2023 09:06:38 GMT
- Title: Discover and Cure: Concept-aware Mitigation of Spurious Correlation
- Authors: Shirley Wu, Mert Yuksekgonul, Linjun Zhang, James Zou
- Abstract summary: Deep neural networks often rely on spurious correlations to make predictions.
We propose an interpretable framework, Discover and Cure (DISC) to tackle the issue.
DISC provides superior generalization ability and interpretability than the existing approaches.
- Score: 14.579651844642616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks often rely on spurious correlations to make predictions,
which hinders generalization beyond training environments. For instance, models
that associate cats with bed backgrounds can fail to predict the existence of
cats in other environments without beds. Mitigating spurious correlations is
crucial in building trustworthy models. However, the existing works lack
transparency to offer insights into the mitigation process. In this work, we
propose an interpretable framework, Discover and Cure (DISC), to tackle the
issue. With human-interpretable concepts, DISC iteratively 1) discovers
unstable concepts across different environments as spurious attributes, then 2)
intervenes on the training data using the discovered concepts to reduce
spurious correlation. Across systematic experiments, DISC provides superior
generalization ability and interpretability than the existing approaches.
Specifically, it outperforms the state-of-the-art methods on an object
recognition task and a skin-lesion classification task by 7.5% and 9.6%,
respectively. Additionally, we offer theoretical analysis and guarantees to
understand the benefits of models trained by DISC. Code and data are available
at https://github.com/Wuyxin/DISC.
Related papers
- Unlearning or Concealment? A Critical Analysis and Evaluation Metrics for Unlearning in Diffusion Models [7.9993879763024065]
We show that the objective functions used for unlearning in the existing methods lead to decoupling of the targeted concepts for the corresponding prompts.
The ineffectiveness of current methods stems primarily from their narrow focus on reducing generation probabilities for specific prompt sets.
We introduce two new evaluation metrics: Concept Retrieval Score (CRS) and Concept Confidence Score (CCS)
arXiv Detail & Related papers (2024-09-09T14:38:31Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - Pre-training and Diagnosing Knowledge Base Completion Models [58.07183284468881]
We introduce and analyze an approach to knowledge transfer from one collection of facts to another without the need for entity or relation matching.
The main contribution is a method that can make use of large-scale pre-training on facts, which were collected from unstructured text.
To understand the obtained pre-trained models better, we then introduce a novel dataset for the analysis of pre-trained models for Open Knowledge Base Completion.
arXiv Detail & Related papers (2024-01-27T15:20:43Z) - COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable
ELements for explaining neural net classifiers on NLP tasks [3.475906200620518]
COCKATIEL is a novel, post-hoc, concept-based, model-agnostic XAI technique.
It generates meaningful explanations from the last layer of a neural net model trained on an NLP classification task.
It does so without compromising the accuracy of the underlying model or requiring a new one to be trained.
arXiv Detail & Related papers (2023-05-11T12:22:20Z) - ESCAPE: Countering Systematic Errors from Machine's Blind Spots via
Interactive Visual Analysis [13.97436974677563]
We propose ESCAPE, a visual analytic system that promotes a human-in-the-loop workflow for countering systematic errors.
By allowing human users to easily inspect spurious associations, the system facilitates users to spontaneously recognize concepts associated misclassifications.
We also propose two statistical approaches, relative concept association to better quantify the associations between a concept and instances, and debias method to mitigate spurious associations.
arXiv Detail & Related papers (2023-03-16T21:29:50Z) - Principled Knowledge Extrapolation with GANs [92.62635018136476]
We study counterfactual synthesis from a new perspective of knowledge extrapolation.
We show that an adversarial game with a closed-form discriminator can be used to address the knowledge extrapolation problem.
Our method enjoys both elegant theoretical guarantees and superior performance in many scenarios.
arXiv Detail & Related papers (2022-05-21T08:39:42Z) - Hybrid Predictive Coding: Inferring, Fast and Slow [62.997667081978825]
We propose a hybrid predictive coding network that combines both iterative and amortized inference in a principled manner.
We demonstrate that our model is inherently sensitive to its uncertainty and adaptively balances balances to obtain accurate beliefs using minimum computational expense.
arXiv Detail & Related papers (2022-04-05T12:52:45Z) - ConCAD: Contrastive Learning-based Cross Attention for Sleep Apnea
Detection [16.938983046369263]
We propose a contrastive learning-based cross attention framework for sleep apnea detection (named ConCAD)
Our proposed framework can be easily integrated into standard deep learning models to utilize expert knowledge and contrastive learning to boost performance.
arXiv Detail & Related papers (2021-05-07T02:38:56Z) - DICE: Diversity in Deep Ensembles via Conditional Redundancy Adversarial
Estimation [109.11580756757611]
Deep ensembles perform better than a single network thanks to the diversity among their members.
Recent approaches regularize predictions to increase diversity; however, they also drastically decrease individual members' performances.
We introduce a novel training criterion called DICE: it increases diversity by reducing spurious correlations among features.
arXiv Detail & Related papers (2021-01-14T10:53:26Z) - Ramifications of Approximate Posterior Inference for Bayesian Deep
Learning in Adversarial and Out-of-Distribution Settings [7.476901945542385]
We show that Bayesian deep learning models on certain occasions marginally outperform conventional neural networks.
Preliminary investigations indicate the potential inherent role of bias due to choices of initialisation, architecture or activation functions.
arXiv Detail & Related papers (2020-09-03T16:58:15Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.