Causal Effect Regularization: Automated Detection and Removal of
Spurious Attributes
- URL: http://arxiv.org/abs/2306.11072v2
- Date: Fri, 8 Dec 2023 01:14:41 GMT
- Title: Causal Effect Regularization: Automated Detection and Removal of
Spurious Attributes
- Authors: Abhinav Kumar, Amit Deshpande, Amit Sharma
- Abstract summary: In many classification datasets, the task labels are spuriously correlated with some input attributes.
We propose a method to automatically identify spurious attributes by estimating their causal effect on the label.
Our method mitigates the reliance on spurious attributes even under noisy estimation of causal effects.
- Score: 13.852987916253685
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many classification datasets, the task labels are spuriously correlated
with some input attributes. Classifiers trained on such datasets often rely on
these attributes for prediction, especially when the spurious correlation is
high, and thus fail to generalize whenever there is a shift in the attributes'
correlation at deployment. If we assume that the spurious attributes are known
a priori, several methods have been proposed to learn a classifier that is
invariant to the specified attributes. However, in real-world data, information
about spurious attributes is typically unavailable. Therefore, we propose a
method to automatically identify spurious attributes by estimating their causal
effect on the label and then use a regularization objective to mitigate the
classifier's reliance on them. Compared to a recent method for identifying
spurious attributes, we find that our method is more accurate in removing the
attribute from the learned model, especially when spurious correlation is high.
Specifically, across synthetic, semi-synthetic, and real-world datasets, our
method shows significant improvement in a metric used to quantify the
dependence of a classifier on spurious attributes ($\Delta$Prob), while
obtaining better or similar accuracy. In addition, our method mitigates the
reliance on spurious attributes even under noisy estimation of causal effects.
To explain the empirical robustness of our method, we create a simple linear
classification task with two sets of attributes: causal and spurious. We prove
that our method only requires that the ranking of estimated causal effects is
correct across attributes to select the correct classifier.
Related papers
- Learning Robust Classifiers with Self-Guided Spurious Correlation Mitigation [26.544938760265136]
Deep neural classifiers rely on spurious correlations between spurious attributes of inputs and targets to make predictions.
We propose a self-guided spurious correlation mitigation framework.
We show that training the classifier to distinguish different prediction behaviors reduces its reliance on spurious correlations without knowing them a priori.
arXiv Detail & Related papers (2024-05-06T17:12:21Z) - A Solution to Co-occurrence Bias: Attributes Disentanglement via Mutual
Information Minimization for Pedestrian Attribute Recognition [10.821982414387525]
We show that current methods can actually suffer in generalizing such fitted attributes interdependencies onto scenes or identities off the dataset distribution.
To render models robust in realistic scenes, we propose the attributes-disentangled feature learning to ensure the recognition of an attribute not inferring on the existence of others.
arXiv Detail & Related papers (2023-07-28T01:34:55Z) - Exploiting Semantic Attributes for Transductive Zero-Shot Learning [97.61371730534258]
Zero-shot learning aims to recognize unseen classes by generalizing the relation between visual features and semantic attributes learned from the seen classes.
We present a novel transductive ZSL method that produces semantic attributes of the unseen data and imposes them on the generative process.
Experiments on five standard benchmarks show that our method yields state-of-the-art results for zero-shot learning.
arXiv Detail & Related papers (2023-03-17T09:09:48Z) - Spread Spurious Attribute: Improving Worst-group Accuracy with Spurious
Attribute Estimation [72.92329724600631]
We propose a pseudo-attribute-based algorithm, coined Spread Spurious Attribute, for improving the worst-group accuracy.
Our experiments on various benchmark datasets show that our algorithm consistently outperforms the baseline methods.
We also demonstrate that the proposed SSA can achieve comparable performances to methods using full (100%) spurious attribute supervision.
arXiv Detail & Related papers (2022-04-05T09:08:30Z) - Semi-FairVAE: Semi-supervised Fair Representation Learning with
Adversarial Variational Autoencoder [92.67156911466397]
We propose a semi-supervised fair representation learning approach based on adversarial variational autoencoder.
We use a bias-aware model to capture inherent bias information on sensitive attribute.
We also use a bias-free model to learn debiased fair representations by using adversarial learning to remove bias information from them.
arXiv Detail & Related papers (2022-04-01T15:57:47Z) - Make an Omelette with Breaking Eggs: Zero-Shot Learning for Novel
Attribute Synthesis [65.74825840440504]
We propose Zero Shot Learning for Attributes (ZSLA), which is the first of its kind to the best of our knowledge.
Our proposed method is able to synthesize the detectors of novel attributes in a zero-shot learning manner.
With using only 32 seen attributes on the Caltech-UCSD Birds-200-2011 dataset, our proposed method is able to synthesize other 207 novel attributes.
arXiv Detail & Related papers (2021-11-28T15:45:54Z) - Evaluating Fairness of Machine Learning Models Under Uncertain and
Incomplete Information [25.739240011015923]
We show that the test accuracy of the attribute classifier is not always correlated with its effectiveness in bias estimation for a downstream model.
Our analysis has surprising and counter-intuitive implications where in certain regimes one might want to distribute the error of the attribute classifier as unevenly as possible.
arXiv Detail & Related papers (2021-02-16T19:02:55Z) - Identifying Spurious Correlations for Robust Text Classification [9.457737910527829]
We propose a method to distinguish spurious and genuine correlations in text classification.
We use features derived from treatment effect estimators to distinguish spurious correlations from "genuine" ones.
Experiments on four datasets suggest that using this approach to inform feature selection also leads to more robust classification.
arXiv Detail & Related papers (2020-10-06T03:49:22Z) - On Irrelevance of Attributes in Flexible Prediction [0.0]
This paper analyses properties of conceptual hierarchy obtained via incremental concept formation method called "flexible prediction"
The impact of selection of simple and combined attributes, of scaling and of distribution of individual attributes and of correlation strengths among them is investigated.
arXiv Detail & Related papers (2020-05-25T08:41:48Z) - Attribute Mix: Semantic Data Augmentation for Fine Grained Recognition [102.45926816660665]
We propose Attribute Mix, a data augmentation strategy at attribute level to expand the fine-grained samples.
The principle lies in that attribute features are shared among fine-grained sub-categories, and can be seamlessly transferred among images.
arXiv Detail & Related papers (2020-04-06T14:06:47Z) - Certified Robustness to Label-Flipping Attacks via Randomized Smoothing [105.91827623768724]
Machine learning algorithms are susceptible to data poisoning attacks.
We present a unifying view of randomized smoothing over arbitrary functions.
We propose a new strategy for building classifiers that are pointwise-certifiably robust to general data poisoning attacks.
arXiv Detail & Related papers (2020-02-07T21:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.