Invariant Structure Learning for Better Generalization and Causal
Explainability
- URL: http://arxiv.org/abs/2206.06469v1
- Date: Mon, 13 Jun 2022 21:04:23 GMT
- Title: Invariant Structure Learning for Better Generalization and Causal
Explainability
- Authors: Yunhao Ge, Sercan \"O. Arik, Jinsung Yoon, Ao Xu, Laurent Itti, Tomas
Pfister
- Abstract summary: We propose a novel framework, Invariant Structure Learning (ISL), to improve causal structure discovery.
ISL splits the data into different environments, and learns a structure that is invariant to the target across different environments.
We demonstrate that ISL accurately discovers the causal structure, outperforms alternative methods, and yields superior generalization for datasets with significant distribution shifts.
- Score: 44.580704853704994
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning the causal structure behind data is invaluable for improving
generalization and obtaining high-quality explanations. We propose a novel
framework, Invariant Structure Learning (ISL), that is designed to improve
causal structure discovery by utilizing generalization as an indication. ISL
splits the data into different environments, and learns a structure that is
invariant to the target across different environments by imposing a consistency
constraint. An aggregation mechanism then selects the optimal classifier based
on a graph structure that reflects the causal mechanisms in the data more
accurately compared to the structures learnt from individual environments.
Furthermore, we extend ISL to a self-supervised learning setting where accurate
causal structure discovery does not rely on any labels. This self-supervised
ISL utilizes invariant causality proposals by iteratively setting different
nodes as targets. On synthetic and real-world datasets, we demonstrate that ISL
accurately discovers the causal structure, outperforms alternative methods, and
yields superior generalization for datasets with significant distribution
shifts.
Related papers
- Dissecting the Failure of Invariant Learning on Graphs [36.11431280689549]
We develop a Structural Causal Model (SCM) to theoretically dissect the performance of two prominent invariant learning methods.
We propose Cross-environment Intra-class Alignment (CIA), which explicitly eliminates spurious features by aligning cross-environment representations conditioned on the same class.
We further propose CIA-LRA (Localized Reweighting Alignment) that leverages the distribution of neighboring labels to selectively align node representations.
arXiv Detail & Related papers (2024-11-05T06:36:48Z) - DeCaf: A Causal Decoupling Framework for OOD Generalization on Node Classification [14.96980804513399]
Graph Neural Networks (GNNs) are susceptible to distribution shifts, creating vulnerability and security issues in critical domains.
Existing methods that target learning an invariant (feature, structure)-label mapping often depend on oversimplified assumptions about the data generation process.
We introduce a more realistic graph data generation model using Structural Causal Models (SCMs)
We propose a casual decoupling framework, DeCaf, that independently learns unbiased feature-label and structure-label mappings.
arXiv Detail & Related papers (2024-10-27T00:22:18Z) - Learning to Model Graph Structural Information on MLPs via Graph Structure Self-Contrasting [50.181824673039436]
We propose a Graph Structure Self-Contrasting (GSSC) framework that learns graph structural information without message passing.
The proposed framework is based purely on Multi-Layer Perceptrons (MLPs), where the structural information is only implicitly incorporated as prior knowledge.
It first applies structural sparsification to remove potentially uninformative or noisy edges in the neighborhood, and then performs structural self-contrasting in the sparsified neighborhood to learn robust node representations.
arXiv Detail & Related papers (2024-09-09T12:56:02Z) - Learning Invariant Representations of Graph Neural Networks via Cluster
Generalization [58.68231635082891]
Graph neural networks (GNNs) have become increasingly popular in modeling graph-structured data.
In this paper, we experimentally find that the performance of GNNs drops significantly when the structure shift happens.
We propose the Cluster Information Transfer (CIT) mechanism, which can learn invariant representations for GNNs.
arXiv Detail & Related papers (2024-03-06T10:36:56Z) - Structured Language Generation Model for Robust Structure Prediction [6.4736137270915215]
We propose a framework that reduces sequence-to-sequence problems to classification problems via methodologies in loss calibration and decoding method.
Our experimental results show that SLGM is able to maintain performance without explicit dataset information, follow and potentially replace dataset-specific fine-tuning.
arXiv Detail & Related papers (2024-02-14T06:33:22Z) - iSCAN: Identifying Causal Mechanism Shifts among Nonlinear Additive
Noise Models [48.33685559041322]
This paper focuses on identifying the causal mechanism shifts in two or more related datasets over the same set of variables.
Code implementing the proposed method is open-source and publicly available at https://github.com/kevinsbello/iSCAN.
arXiv Detail & Related papers (2023-06-30T01:48:11Z) - SE-GSL: A General and Effective Graph Structure Learning Framework
through Structural Entropy Optimization [67.28453445927825]
Graph Neural Networks (GNNs) are de facto solutions to structural data learning.
Existing graph structure learning (GSL) frameworks still lack robustness and interpretability.
This paper proposes a general GSL framework, SE-GSL, through structural entropy and the graph hierarchy abstracted in the encoding tree.
arXiv Detail & Related papers (2023-03-17T05:20:24Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Structure-Aware Feature Generation for Zero-Shot Learning [108.76968151682621]
We introduce a novel structure-aware feature generation scheme, termed as SA-GAN, to account for the topological structure in learning both the latent space and the generative networks.
Our method significantly enhances the generalization capability on unseen-classes and consequently improve the classification performance.
arXiv Detail & Related papers (2021-08-16T11:52:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.