When Does Group Invariant Learning Survive Spurious Correlations?
- URL: http://arxiv.org/abs/2206.14534v1
- Date: Wed, 29 Jun 2022 11:16:11 GMT
- Title: When Does Group Invariant Learning Survive Spurious Correlations?
- Authors: Yimeng Chen, Ruibin Xiong, Zhiming Ma, Yanyan Lan
- Abstract summary: In this paper, we reveal the insufficiency of existing group invariant learning methods.
We propose two criteria on judging such sufficiency.
We show that existing methods can violate both criteria and thus fail in generalizing to spurious correlation shifts.
Motivated by this, we design a new group invariant learning method, which constructs groups with statistical independence tests.
- Score: 29.750875769713513
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: By inferring latent groups in the training data, recent works introduce
invariant learning to the case where environment annotations are unavailable.
Typically, learning group invariance under a majority/minority split is
empirically shown to be effective in improving out-of-distribution
generalization on many datasets. However, theoretical guarantee for these
methods on learning invariant mechanisms is lacking. In this paper, we reveal
the insufficiency of existing group invariant learning methods in preventing
classifiers from depending on spurious correlations in the training set.
Specifically, we propose two criteria on judging such sufficiency.
Theoretically and empirically, we show that existing methods can violate both
criteria and thus fail in generalizing to spurious correlation shifts.
Motivated by this, we design a new group invariant learning method, which
constructs groups with statistical independence tests, and reweights samples by
group label proportion to meet the criteria. Experiments on both synthetic and
real data demonstrate that the new method significantly outperforms existing
group invariant learning methods in generalizing to spurious correlation
shifts.
Related papers
- Trained Models Tell Us How to Make Them Robust to Spurious Correlation without Group Annotation [3.894771553698554]
Empirical Risk Minimization (ERM) models tend to rely on attributes that have high spurious correlation with the target.
This can degrade the performance on underrepresented (or'minority') groups that lack these attributes.
We propose Environment-based Validation and Loss-based Sampling (EVaLS) to enhance robustness to spurious correlation.
arXiv Detail & Related papers (2024-10-07T08:17:44Z) - Learning with Complementary Labels Revisited: The Selected-Completely-at-Random Setting Is More Practical [66.57396042747706]
Complementary-label learning is a weakly supervised learning problem.
We propose a consistent approach that does not rely on the uniform distribution assumption.
We find that complementary-label learning can be expressed as a set of negative-unlabeled binary classification problems.
arXiv Detail & Related papers (2023-11-27T02:59:17Z) - On The Impact of Machine Learning Randomness on Group Fairness [11.747264308336012]
We investigate the impact on group fairness of different sources of randomness in training neural networks.
We show that the variance in group fairness measures is rooted in the high volatility of the learning process on under-represented groups.
We show how one can control group-level accuracy, with high efficiency and negligible impact on the model's overall performance, by simply changing the data order for a single epoch.
arXiv Detail & Related papers (2023-07-09T09:36:31Z) - Adaptive Negative Evidential Deep Learning for Open-set Semi-supervised Learning [69.81438976273866]
Open-set semi-supervised learning (Open-set SSL) considers a more practical scenario, where unlabeled data and test data contain new categories (outliers) not observed in labeled data (inliers)
We introduce evidential deep learning (EDL) as an outlier detector to quantify different types of uncertainty, and design different uncertainty metrics for self-training and inference.
We propose a novel adaptive negative optimization strategy, making EDL more tailored to the unlabeled dataset containing both inliers and outliers.
arXiv Detail & Related papers (2023-03-21T09:07:15Z) - Leveraging Structure for Improved Classification of Grouped Biased Data [8.121462458089143]
We consider semi-supervised binary classification for applications in which data points are naturally grouped.
We derive a semi-supervised algorithm that explicitly leverages the structure to learn an optimal, group-aware, probability-outputd classifier.
arXiv Detail & Related papers (2022-12-07T15:18:21Z) - Outlier-Robust Group Inference via Gradient Space Clustering [50.87474101594732]
Existing methods can improve the worst-group performance, but they require group annotations, which are often expensive and sometimes infeasible to obtain.
We address the problem of learning group annotations in the presence of outliers by clustering the data in the space of gradients of the model parameters.
We show that data in the gradient space has a simpler structure while preserving information about minority groups and outliers, making it suitable for standard clustering methods like DBSCAN.
arXiv Detail & Related papers (2022-10-13T06:04:43Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - Heterogeneous Risk Minimization [25.5458915855661]
Invariant learning methods for out-of-distribution generalization have been proposed by leveraging multiple training environments to find invariant relationships.
Modern datasets are assembled by merging data from multiple sources without explicit source labels.
We propose Heterogeneous Risk Minimization (HRM) framework to achieve joint learning of latent heterogeneity among the data and invariant relationship.
arXiv Detail & Related papers (2021-05-09T02:51:36Z) - Neural Networks for Learning Counterfactual G-Invariances from Single
Environments [13.848760376470038]
neural networks are believed to have difficulties extrapolating beyond training data distribution.
This work shows that, for extrapolations based on finite transformation groups, a model's inability to extrapolate is unrelated to its capacity.
arXiv Detail & Related papers (2021-04-20T16:35:35Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.