Mind the Label Shift of Augmentation-based Graph OOD Generalization
- URL: http://arxiv.org/abs/2303.14859v1
- Date: Mon, 27 Mar 2023 00:08:45 GMT
- Title: Mind the Label Shift of Augmentation-based Graph OOD Generalization
- Authors: Junchi Yu and Jian Liang and Ran He
- Abstract summary: LiSA exploits textbfLabel-textbfinvariant textbfSubgraphs of the training graphs to construct textbfAugmented environments.
LiSA generates diverse augmented environments with a consistent predictive relationship.
Experiments on node-level and graph-level OOD benchmarks show that LiSA achieves impressive generalization performance with different GNN backbones.
- Score: 88.32356432272356
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Out-of-distribution (OOD) generalization is an important issue for Graph
Neural Networks (GNNs). Recent works employ different graph editions to
generate augmented environments and learn an invariant GNN for generalization.
However, the label shift usually occurs in augmentation since graph structural
edition inevitably alters the graph label. This brings inconsistent predictive
relationships among augmented environments, which is harmful to generalization.
To address this issue, we propose \textbf{LiSA}, which generates
label-invariant augmentations to facilitate graph OOD generalization. Instead
of resorting to graph editions, LiSA exploits \textbf{L}abel-\textbf{i}nvariant
\textbf{S}ubgraphs of the training graphs to construct \textbf{A}ugmented
environments. Specifically, LiSA first designs the variational subgraph
generators to extract locally predictive patterns and construct multiple
label-invariant subgraphs efficiently. Then, the subgraphs produced by
different generators are collected to build different augmented environments.
To promote diversity among augmented environments, LiSA further introduces a
tractable energy-based regularization to enlarge pair-wise distances between
the distributions of environments. In this manner, LiSA generates diverse
augmented environments with a consistent predictive relationship and
facilitates learning an invariant GNN. Extensive experiments on node-level and
graph-level OOD benchmarks show that LiSA achieves impressive generalization
performance with different GNN backbones. Code is available on
\url{https://github.com/Samyu0304/LiSA}.
Related papers
- Subgraph Aggregation for Out-of-Distribution Generalization on Graphs [29.884717215947745]
Out-of-distribution (OOD) generalization in Graph Neural Networks (GNNs) has gained significant attention.
We propose a novel framework, SubGraph Aggregation (SuGAr), designed to learn a diverse set of subgraphs.
Experiments on both synthetic and real-world datasets demonstrate that SuGAr outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-10-29T16:54:37Z) - Improving out-of-distribution generalization in graphs via hierarchical semantic environments [5.481047026874547]
We propose a novel approach to generate hierarchical environments for each graph.
We introduce a new learning objective that guides our model to learn the diversity of environments within the same hierarchy.
Our framework achieves up to 1.29% and 2.83% improvement over the best baselines on IC50 and EC50 prediction tasks, respectively.
arXiv Detail & Related papers (2024-03-04T07:03:10Z) - Graph Out-of-Distribution Generalization via Causal Intervention [69.70137479660113]
We introduce a conceptually simple yet principled approach for training robust graph neural networks (GNNs) under node-level distribution shifts.
Our method resorts to a new learning objective derived from causal inference that coordinates an environment estimator and a mixture-of-expert GNN predictor.
Our model can effectively enhance generalization with various types of distribution shifts and yield up to 27.4% accuracy improvement over state-of-the-arts on graph OOD generalization benchmarks.
arXiv Detail & Related papers (2024-02-18T07:49:22Z) - GSINA: Improving Subgraph Extraction for Graph Invariant Learning via
Graph Sinkhorn Attention [52.67633391931959]
Graph invariant learning (GIL) has been an effective approach to discovering the invariant relationships between graph data and its labels.
We propose a novel graph attention mechanism called Graph Sinkhorn Attention (GSINA)
GSINA is able to obtain meaningful, differentiable invariant subgraphs with controllable sparsity and softness.
arXiv Detail & Related papers (2024-02-11T12:57:16Z) - Graph Invariant Learning with Subgraph Co-mixup for Out-Of-Distribution
Generalization [51.913685334368104]
We propose a novel graph invariant learning method based on invariant and variant patterns co-mixup strategy.
Our method significantly outperforms state-of-the-art under various distribution shifts.
arXiv Detail & Related papers (2023-12-18T07:26:56Z) - Graph Mixture of Experts: Learning on Large-Scale Graphs with Explicit
Diversity Modeling [60.0185734837814]
Graph neural networks (GNNs) have found extensive applications in learning from graph data.
To bolster the generalization capacity of GNNs, it has become customary to augment training graph structures with techniques like graph augmentations.
This study introduces the concept of Mixture-of-Experts (MoE) to GNNs, with the aim of augmenting their capacity to adapt to a diverse range of training graph structures.
arXiv Detail & Related papers (2023-04-06T01:09:36Z) - Hybrid Augmented Automated Graph Contrastive Learning [3.785553471764994]
We propose a framework called Hybrid Augmented Automated Graph Contrastive Learning (HAGCL)
HAGCL consists of a feature-level learnable view generator and an edge-level learnable view generator.
It insures to learn the most semantically meaningful structure in terms of features and topology.
arXiv Detail & Related papers (2023-03-24T03:26:20Z) - Finding Diverse and Predictable Subgraphs for Graph Domain
Generalization [88.32356432272356]
This paper focuses on out-of-distribution generalization on graphs where performance drops due to the unseen distribution shift.
We propose a new graph domain generalization framework, dubbed as DPS, by constructing multiple populations from the source domains.
Experiments on both node-level and graph-level benchmarks shows that the proposed DPS achieves impressive performance for various graph domain generalization tasks.
arXiv Detail & Related papers (2022-06-19T07:57:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.