Graph Structure Learning with Variational Information Bottleneck
- URL: http://arxiv.org/abs/2112.08903v1
- Date: Thu, 16 Dec 2021 14:22:13 GMT
- Title: Graph Structure Learning with Variational Information Bottleneck
- Authors: Qingyun Sun, Jianxin Li, Hao Peng, Jia Wu, Xingcheng Fu, Cheng Ji,
Philip S. Yu
- Abstract summary: We propose a novel Variational Information Bottleneck guided Graph Structure Learning framework, namely VIB-GSL.
VIB-GSL learns an informative and compressive graph structure to distill the actionable information for specific downstream tasks.
- Score: 70.62851953251253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Neural Networks (GNNs) have shown promising results on a broad spectrum
of applications. Most empirical studies of GNNs directly take the observed
graph as input, assuming the observed structure perfectly depicts the accurate
and complete relations between nodes. However, graphs in the real world are
inevitably noisy or incomplete, which could even exacerbate the quality of
graph representations. In this work, we propose a novel Variational Information
Bottleneck guided Graph Structure Learning framework, namely VIB-GSL, in the
perspective of information theory. VIB-GSL advances the Information Bottleneck
(IB) principle for graph structure learning, providing a more elegant and
universal framework for mining underlying task-relevant relations. VIB-GSL
learns an informative and compressive graph structure to distill the actionable
information for specific downstream tasks. VIB-GSL deduces a variational
approximation for irregular graph data to form a tractable IB objective
function, which facilitates training stability. Extensive experimental results
demonstrate that the superior effectiveness and robustness of VIB-GSL.
Related papers
- Robust Graph Structure Learning with the Alignment of Features and
Adjacency Matrix [8.711977569042865]
Many approaches have been proposed for graph structure learning (GSL) to jointly learn a clean graph structure and corresponding representations.
This paper proposes a novel regularized GSL approach, particularly with an alignment of feature information and graph information.
We conduct experiments on real-world graphs to evaluate the effectiveness of our approach.
arXiv Detail & Related papers (2023-07-05T09:05:14Z) - Learning Strong Graph Neural Networks with Weak Information [64.64996100343602]
We develop a principled approach to the problem of graph learning with weak information (GLWI)
We propose D$2$PT, a dual-channel GNN framework that performs long-range information propagation on the input graph with incomplete structure, but also on a global graph that encodes global semantic similarities.
arXiv Detail & Related papers (2023-05-29T04:51:09Z) - Towards Unsupervised Deep Graph Structure Learning [67.58720734177325]
We propose an unsupervised graph structure learning paradigm, where the learned graph topology is optimized by data itself without any external guidance.
Specifically, we generate a learning target from the original data as an "anchor graph", and use a contrastive loss to maximize the agreement between the anchor graph and the learned graph.
arXiv Detail & Related papers (2022-01-17T11:57:29Z) - Improving Subgraph Recognition with Variational Graph Information
Bottleneck [62.69606854404757]
Subgraph recognition aims at discovering a compressed substructure of a graph that is most informative to the graph property.
This paper introduces a noise injection method to compress the information in the subgraphs, which leads to a novel Variational Graph Information Bottleneck (VGIB) framework.
arXiv Detail & Related papers (2021-12-18T10:51:13Z) - Recognizing Predictive Substructures with Subgraph Information
Bottleneck [97.19131149357234]
We propose a novel subgraph information bottleneck (SIB) framework to recognize such subgraphs, named IB-subgraph.
Intractability of mutual information and the discrete nature of graph data makes the objective of SIB notoriously hard to optimize.
Experiments on graph learning and large-scale point cloud tasks demonstrate the superior property of IB-subgraph.
arXiv Detail & Related papers (2021-03-20T11:19:43Z) - Deep Graph Structure Learning for Robust Representations: A Survey [20.564611153151834]
Graph Neural Networks (GNNs) are widely used for analyzing graph-structured data.
To improve the robustness of GNN models, many studies have been proposed around the central concept of Graph Structure Learning.
arXiv Detail & Related papers (2021-03-04T13:49:25Z) - Graph Information Bottleneck [77.21967740646784]
Graph Neural Networks (GNNs) provide an expressive way to fuse information from network structure and node features.
Inheriting from the general Information Bottleneck (IB), GIB aims to learn the minimal sufficient representation for a given task.
We show that our proposed models are more robust than state-of-the-art graph defense models.
arXiv Detail & Related papers (2020-10-24T07:13:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.