Conditional entropy minimization principle for learning domain invariant
representation features
- URL: http://arxiv.org/abs/2201.10460v1
- Date: Tue, 25 Jan 2022 17:02:12 GMT
- Title: Conditional entropy minimization principle for learning domain invariant
representation features
- Authors: Thuan Nguyen, Boyang Lyu, Prakash Ishwar, Matthias Scheutz, Shuchin
Aeron
- Abstract summary: In this paper, we propose a framework based on the conditional entropy minimization principle to filter out the spurious invariant features.
We show that the proposed approach is closely related to the well-known Information Bottleneck framework.
- Score: 30.459247038765568
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Invariance principle-based methods, for example, Invariant Risk Minimization
(IRM), have recently emerged as promising approaches for Domain Generalization
(DG). Despite the promising theory, invariance principle-based approaches fail
in common classification tasks due to the mixture of the true invariant
features and the spurious invariant features. In this paper, we propose a
framework based on the conditional entropy minimization principle to filter out
the spurious invariant features leading to a new algorithm with a better
generalization capability. We theoretically prove that under some particular
assumptions, the representation function can precisely recover the true
invariant features. In addition, we also show that the proposed approach is
closely related to the well-known Information Bottleneck framework. Both the
theoretical and numerical results are provided to justify our approach.
Related papers
- A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - Generalized Invariant Matching Property via LASSO [19.786769414376323]
In this work, we generalize the invariant matching property by formulating a high-dimensional problem with intrinsic sparsity.
We propose a more robust and computation-efficient algorithm by leveraging a variant of Lasso.
arXiv Detail & Related papers (2023-01-14T21:09:30Z) - Variational Disentanglement for Domain Generalization [68.85458536180437]
We propose to tackle the problem of domain generalization by delivering an effective framework named Variational Disentanglement Network (VDN)
VDN is capable of disentangling the domain-specific features and task-specific features, where the task-specific features are expected to be better generalized to unseen but related test data.
arXiv Detail & Related papers (2021-09-13T09:55:32Z) - Invariance Principle Meets Information Bottleneck for
Out-of-Distribution Generalization [77.24152933825238]
We show that for linear classification tasks we need stronger restrictions on the distribution shifts, or otherwise OOD generalization is impossible.
We prove that a form of the information bottleneck constraint along with invariance helps address key failures when invariant features capture all the information about the label and also retains the existing success when they do not.
arXiv Detail & Related papers (2021-06-11T20:42:27Z) - On the benefits of representation regularization in invariance based
domain generalization [6.197602794925773]
Domain generalization aims to alleviate such a prediction gap between the observed and unseen environments.
In this paper, we reveal that merely learning invariant representation is vulnerable to the unseen environment.
Our analysis further inspires an efficient regularization method to improve the robustness in domain generalization.
arXiv Detail & Related papers (2021-05-30T13:13:55Z) - Statistical optimality and stability of tangent transform algorithms in
logit models [6.9827388859232045]
We provide conditions on the data generating process to derive non-asymptotic upper bounds to the risk incurred by the logistical optima.
In particular, we establish local variation of the algorithm without any assumptions on the data-generating process.
We explore a special case involving a semi-orthogonal design under which a global convergence is obtained.
arXiv Detail & Related papers (2020-10-25T05:15:13Z) - The Risks of Invariant Risk Minimization [52.7137956951533]
Invariant Risk Minimization is an objective based on the idea for learning deep, invariant features of data.
We present the first analysis of classification under the IRM objective--as well as these recently proposed alternatives--under a fairly natural and general model.
We show that IRM can fail catastrophically unless the test data are sufficiently similar to the training distribution--this is precisely the issue that it was intended to solve.
arXiv Detail & Related papers (2020-10-12T14:54:32Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - Learning explanations that are hard to vary [75.30552491694066]
We show that averaging across examples can favor memorization and patchwork' solutions that sew together different strategies.
We then propose and experimentally validate a simple alternative algorithm based on a logical AND.
arXiv Detail & Related papers (2020-09-01T10:17:48Z) - Convex Representation Learning for Generalized Invariance in
Semi-Inner-Product Space [32.442549424823355]
In this work we develop an algorithm for a variety of generalized representations in a semi-norms that representers in a lead, and bounds are established.
This allows in representations to be learned efficiently and effectively as confirmed in our experiments along with accurate predictions.
arXiv Detail & Related papers (2020-04-25T18:54:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.