Counterfactual Supervision-based Information Bottleneck for
Out-of-Distribution Generalization
- URL: http://arxiv.org/abs/2208.07798v1
- Date: Tue, 16 Aug 2022 15:26:00 GMT
- Title: Counterfactual Supervision-based Information Bottleneck for
Out-of-Distribution Generalization
- Authors: Bin Deng and Kui Jia
- Abstract summary: We show that the invariant risk minimization algorithm (IB-IRM) is not sufficient for learning invariant features in linear classification problems.
We propose a textitCounterfactual Supervision-based Information Bottleneck (CSIB) learning algorithm that provably recovers the invariant features.
- Score: 40.94431121318241
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning invariant (causal) features for out-of-distribution (OOD)
generalization has attracted extensive attention recently, and among the
proposals invariant risk minimization (IRM) (Arjovsky et al., 2019) is a
notable solution. In spite of its theoretical promise for linear regression,
the challenges of using IRM in linear classification problems yet remain
(Rosenfeld et al.,2020, Nagarajan et al., 2021). Along this line, a recent
study (Arjovsky et al., 2019) has made a first step and proposes a learning
principle of information bottleneck based invariant risk minimization (IB-IRM).
In this paper, we first show that the key assumption of support overlap of
invariant features used in (Arjovsky et al., 2019) is rather strong for the
guarantee of OOD generalization and it is still possible to achieve the optimal
solution without such assumption. To further answer the question of whether
IB-IRM is sufficient for learning invariant features in linear classification
problems, we show that IB-IRM would still fail in two cases whether or not the
invariant features capture all information about the label. To address such
failures, we propose a \textit{Counterfactual Supervision-based Information
Bottleneck (CSIB)} learning algorithm that provably recovers the invariant
features. The proposed algorithm works even when accessing data from a single
environment, and has theoretically consistent results for both binary and
multi-class problems. We present empirical experiments on three synthetic
datasets that verify the efficacy of our proposed method.
Related papers
- Invariant-Feature Subspace Recovery: A New Class of Provable Domain
Generalization Algorithms [14.248005245508432]
Domain generalization asks for trained models over a set of training environments to generalize well in unseen test environments.
We propose Subspace Recovery (ISR): a new class of algorithms to achieve provable regression problems.
ISR can be used as post-processing methods for neural nets such as neural nets Empirically, we demonstrate the superior performance of our ISRs on synthetic benchmarks.
arXiv Detail & Related papers (2023-11-02T03:24:55Z) - Coping with Change: Learning Invariant and Minimum Sufficient
Representations for Fine-Grained Visual Categorization [26.254072665916155]
Fine-grained visual categorization (FGVC) is a challenging task due to similar visual appearances between various species.
Previous studies assume that the training and test data have the same underlying distributions, and that features extracted by modern backbone architectures remain discriminative and generalize well to unseen test data.
We combine the merits of invariant risk minimization (IRM) and information bottleneck (IB) principle to learn invariant and minimum sufficient (IMS) representations for FGVC.
arXiv Detail & Related papers (2023-06-08T02:45:15Z) - Mutual Information Learned Regressor: an Information-theoretic Viewpoint
of Training Regression Systems [10.314518385506007]
An existing common practice for solving regression problems is the mean square error (MSE) minimization approach.
Recently, Yi et al., proposed a mutual information based supervised learning framework where they introduced a label entropy regularization.
In this paper, we investigate the regression under the mutual information based supervised learning framework.
arXiv Detail & Related papers (2022-11-23T03:43:22Z) - GEC: A Unified Framework for Interactive Decision Making in MDP, POMDP,
and Beyond [101.5329678997916]
We study sample efficient reinforcement learning (RL) under the general framework of interactive decision making.
We propose a novel complexity measure, generalized eluder coefficient (GEC), which characterizes the fundamental tradeoff between exploration and exploitation.
We show that RL problems with low GEC form a remarkably rich class, which subsumes low Bellman eluder dimension problems, bilinear class, low witness rank problems, PO-bilinear class, and generalized regular PSR.
arXiv Detail & Related papers (2022-11-03T16:42:40Z) - Kernelized Heterogeneous Risk Minimization [25.5458915855661]
We propose a Kernelized Heterogeneous Risk Minimization (KerHRM) algorithm, which achieves both the latent exploration and invariant learning in kernel space.
We theoretically justify our algorithm and empirically validate the effectiveness of our algorithm with extensive experiments.
arXiv Detail & Related papers (2021-10-24T12:26:50Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - The Risks of Invariant Risk Minimization [52.7137956951533]
Invariant Risk Minimization is an objective based on the idea for learning deep, invariant features of data.
We present the first analysis of classification under the IRM objective--as well as these recently proposed alternatives--under a fairly natural and general model.
We show that IRM can fail catastrophically unless the test data are sufficiently similar to the training distribution--this is precisely the issue that it was intended to solve.
arXiv Detail & Related papers (2020-10-12T14:54:32Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - A Generic First-Order Algorithmic Framework for Bi-Level Programming
Beyond Lower-Level Singleton [49.23948907229656]
Bi-level Descent Aggregation is a flexible and modularized algorithmic framework for generic bi-level optimization.
We derive a new methodology to prove the convergence of BDA without the LLS condition.
Our investigations also demonstrate that BDA is indeed compatible to a verify of particular first-order computation modules.
arXiv Detail & Related papers (2020-06-07T05:18:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.