Invariant and Transportable Representations for Anti-Causal Domain
Shifts
- URL: http://arxiv.org/abs/2207.01603v1
- Date: Mon, 4 Jul 2022 17:36:49 GMT
- Title: Invariant and Transportable Representations for Anti-Causal Domain
Shifts
- Authors: Yibo Jiang, Victor Veitch
- Abstract summary: We show how to leverage the shared causal structure of the domains to learn a representation that both admits an invariant predictor and that allows fast adaptation in new domains.
Experiments on both synthetic and real-world data demonstrate the effectiveness of the proposed learning algorithm.
- Score: 18.530198688722752
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world classification problems must contend with domain shift, the
(potential) mismatch between the domain where a model is deployed and the
domain(s) where the training data was gathered. Methods to handle such problems
must specify what structure is common between the domains and what varies. A
natural assumption is that causal (structural) relationships are invariant in
all domains. Then, it is tempting to learn a predictor for label $Y$ that
depends only on its causal parents. However, many real-world problems are
"anti-causal" in the sense that $Y$ is a cause of the covariates $X$ -- in this
case, $Y$ has no causal parents and the naive causal invariance is useless. In
this paper, we study representation learning under a particular notion of
domain shift that both respects causal invariance and that naturally handles
the "anti-causal" structure. We show how to leverage the shared causal
structure of the domains to learn a representation that both admits an
invariant predictor and that also allows fast adaptation in new domains. The
key is to translate causal assumptions into learning principles that
disentangle "invariant" and "non-stable" features. Experiments on both
synthetic and real-world data demonstrate the effectiveness of the proposed
learning algorithm. Code is available at https://github.com/ybjiaang/ACTIR.
Related papers
- Domain Generalization via Causal Adjustment for Cross-Domain Sentiment
Analysis [59.73582306457387]
We focus on the problem of domain generalization for cross-domain sentiment analysis.
We propose a backdoor adjustment-based causal model to disentangle the domain-specific and domain-invariant representations.
A series of experiments show the great performance and robustness of our model.
arXiv Detail & Related papers (2024-02-22T13:26:56Z) - Variational Counterfactual Prediction under Runtime Domain Corruption [50.89405221574912]
Co-occurrence of domain shift and inaccessible variables runtime domain corruption seriously impairs generalizability of trained counterfactual predictor.
We build an adversarially unified variational causal effect model, named VEGAN, with a novel two-stage adversarial domain adaptation scheme.
We demonstrate that VEGAN outperforms other state-of-the-art baselines on individual-level treatment effect estimation in the presence of runtime domain corruption.
arXiv Detail & Related papers (2023-06-23T02:54:34Z) - Improving Domain Generalization with Domain Relations [77.63345406973097]
This paper focuses on domain shifts, which occur when the model is applied to new domains that are different from the ones it was trained on.
We propose a new approach called D$3$G to learn domain-specific models.
Our results show that D$3$G consistently outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-02-06T08:11:16Z) - A Unified Causal View of Domain Invariant Representation Learning [19.197022592928164]
Machine learning methods can be unreliable when deployed in domains that differ from the domains on which they were trained.
This paper shows how the different methods relate to each other, and clarify the real-world circumstances under which each is expected to succeed.
arXiv Detail & Related papers (2022-08-15T03:08:58Z) - Causality Inspired Representation Learning for Domain Generalization [47.574964496891404]
We introduce a general structural causal model to formalize the Domain generalization problem.
Our goal is to extract the causal factors from inputs and then reconstruct the invariant causal mechanisms.
We highlight that ideal causal factors should meet three basic properties: separated from the non-causal ones, jointly independent, and causally sufficient for the classification.
arXiv Detail & Related papers (2022-03-27T08:08:33Z) - Domain-Class Correlation Decomposition for Generalizable Person
Re-Identification [34.813965300584776]
In person re-identification, the domain and class are correlated.
We show that domain adversarial learning will lose certain information about class due to this domain-class correlation.
Our model outperforms the state-of-the-art methods on the large-scale domain generalization Re-ID benchmark.
arXiv Detail & Related papers (2021-06-29T09:45:03Z) - Learning Domain Invariant Representations by Joint Wasserstein Distance
Minimization [3.382067152367334]
Domain shifts in the training data are common in practical applications of machine learning.
Ideally, a ML model should work well independently of these shifts, for example, by learning a domain-invariant representation.
Common ML losses do not give strong guarantees on how consistently the ML model performs for different domains.
arXiv Detail & Related papers (2021-06-09T09:08:51Z) - Counterfactual Invariance to Spurious Correlations: Why and How to Pass
Stress Tests [87.60900567941428]
A spurious correlation' is the dependence of a model on some aspect of the input data that an analyst thinks shouldn't matter.
In machine learning, these have a know-it-when-you-see-it character.
We study stress testing using the tools of causal inference.
arXiv Detail & Related papers (2021-05-31T14:39:38Z) - A Bit More Bayesian: Domain-Invariant Learning with Uncertainty [111.22588110362705]
Domain generalization is challenging due to the domain shift and the uncertainty caused by the inaccessibility of target domain data.
In this paper, we address both challenges with a probabilistic framework based on variational Bayesian inference.
We derive domain-invariant representations and classifiers, which are jointly established in a two-layer Bayesian neural network.
arXiv Detail & Related papers (2021-05-09T21:33:27Z) - Heuristic Domain Adaptation [105.59792285047536]
Heuristic Domain Adaptation Network (HDAN) explicitly learns the domain-invariant and domain-specific representations.
Heuristic Domain Adaptation Network (HDAN) has exceeded state-of-the-art on unsupervised DA, multi-source DA and semi-supervised DA.
arXiv Detail & Related papers (2020-11-30T04:21:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.