Unifying Invariance and Spuriousity for Graph Out-of-Distribution via
Probability of Necessity and Sufficiency
- URL: http://arxiv.org/abs/2402.09165v1
- Date: Wed, 14 Feb 2024 13:31:53 GMT
- Title: Unifying Invariance and Spuriousity for Graph Out-of-Distribution via
Probability of Necessity and Sufficiency
- Authors: Xuexin Chen, Ruichu Cai, Kaitao Zheng, Zhifan Jiang, Zhengting Huang,
Zhifeng Hao, Zijian Li
- Abstract summary: We propose a unified framework to exploit the Probability of Necessity and Sufficiency to extract the Invariant Substructure (PNSIS)
Our model outperforms the state-of-the-art techniques on graph OOD on several benchmarks.
- Score: 19.49531172542614
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Out-of-Distribution (OOD), requiring that models trained on biased data
generalize to the unseen test data, has a massive of real-world applications.
One of the most mainstream methods is to extract the invariant subgraph by
aligning the original and augmented data with the help of environment
augmentation. However, these solutions might lead to the loss or redundancy of
semantic subgraph and further result in suboptimal generalization. To address
this challenge, we propose a unified framework to exploit the Probability of
Necessity and Sufficiency to extract the Invariant Substructure (PNSIS). Beyond
that, this framework further leverages the spurious subgraph to boost the
generalization performance in an ensemble manner to enhance the robustness on
the noise data. Specificially, we first consider the data generation process
for graph data. Under mild conditions, we show that the invariant subgraph can
be extracted by minimizing an upper bound, which is built on the theoretical
advance of probability of necessity and sufficiency. To further bridge the
theory and algorithm, we devise the PNSIS model, which involves an invariant
subgraph extractor for invariant graph learning as well invariant and spurious
subgraph classifiers for generalization enhancement. Experimental results
demonstrate that our \textbf{PNSIS} model outperforms the state-of-the-art
techniques on graph OOD on several benchmarks, highlighting the effectiveness
in real-world scenarios.
Related papers
- Subgraph Aggregation for Out-of-Distribution Generalization on Graphs [29.884717215947745]
Out-of-distribution (OOD) generalization in Graph Neural Networks (GNNs) has gained significant attention.
We propose a novel framework, SubGraph Aggregation (SuGAr), designed to learn a diverse set of subgraphs.
Experiments on both synthetic and real-world datasets demonstrate that SuGAr outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-10-29T16:54:37Z) - Sub-graph Based Diffusion Model for Link Prediction [43.15741675617231]
Denoising Diffusion Probabilistic Models (DDPMs) represent a contemporary class of generative models with exceptional qualities.
We build a novel generative model for link prediction using a dedicated design to decompose the likelihood estimation process via the Bayesian formula.
Our proposed method presents numerous advantages: (1) transferability across datasets without retraining, (2) promising generalization on limited training data, and (3) robustness against graph adversarial attacks.
arXiv Detail & Related papers (2024-09-13T02:23:55Z) - DIVE: Subgraph Disagreement for Graph Out-of-Distribution Generalization [44.291382840373]
This paper addresses the challenge of out-of-distribution generalization in graph machine learning.
Traditional graph learning algorithms falter in real-world scenarios where this assumption fails.
A principal factor contributing to this suboptimal performance is the inherent simplicity bias of neural networks.
arXiv Detail & Related papers (2024-08-08T12:08:55Z) - Unifying Invariant and Variant Features for Graph Out-of-Distribution via Probability of Necessity and Sufficiency [18.564387153282293]
We propose exploiting Probability of Necessity and Sufficiency (PNS) to extract sufficient and necessary invariant substructures.
We also leverage the domain variant subgraphs related to the labels to boost the generalization performance in an ensemble manner.
Experimental results demonstrate that our SNIGL model outperforms the state-of-the-art techniques on six public benchmarks.
arXiv Detail & Related papers (2024-07-21T21:35:01Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Invariance Principle Meets Out-of-Distribution Generalization on Graphs [66.04137805277632]
Complex nature of graphs thwarts the adoption of the invariance principle for OOD generalization.
domain or environment partitions, which are often required by OOD methods, can be expensive to obtain for graphs.
We propose a novel framework to explicitly model this process using a contrastive strategy.
arXiv Detail & Related papers (2022-02-11T04:38:39Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Recognizing Predictive Substructures with Subgraph Information
Bottleneck [97.19131149357234]
We propose a novel subgraph information bottleneck (SIB) framework to recognize such subgraphs, named IB-subgraph.
Intractability of mutual information and the discrete nature of graph data makes the objective of SIB notoriously hard to optimize.
Experiments on graph learning and large-scale point cloud tasks demonstrate the superior property of IB-subgraph.
arXiv Detail & Related papers (2021-03-20T11:19:43Z) - Graph Information Bottleneck for Subgraph Recognition [103.37499715761784]
We propose a framework of Graph Information Bottleneck (GIB) for the subgraph recognition problem in deep graph learning.
Under this framework, one can recognize the maximally informative yet compressive subgraph, named IB-subgraph.
We evaluate the properties of the IB-subgraph in three application scenarios: improvement of graph classification, graph interpretation and graph denoising.
arXiv Detail & Related papers (2020-10-12T09:32:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.