Multi-View Graph Representation Learning Beyond Homophily
- URL: http://arxiv.org/abs/2304.07509v1
- Date: Sat, 15 Apr 2023 08:35:49 GMT
- Title: Multi-View Graph Representation Learning Beyond Homophily
- Authors: Bei Lin, You Li, Ning Gui, Zhuopeng Xu, Zhiwu Yu
- Abstract summary: Unsupervised graph representation learning(GRL) aims to distill diverse graph information into task-agnostic embeddings without label supervision.
A novel framework, denoted as Multi-view Graph(MVGE) is proposed, and a set of key designs are identified.
- Score: 2.601278669926709
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unsupervised graph representation learning(GRL) aims to distill diverse graph
information into task-agnostic embeddings without label supervision. Due to a
lack of support from labels, recent representation learning methods usually
adopt self-supervised learning, and embeddings are learned by solving a
handcrafted auxiliary task(so-called pretext task). However, partially due to
the irregular non-Euclidean data in graphs, the pretext tasks are generally
designed under homophily assumptions and cornered in the low-frequency signals,
which results in significant loss of other signals, especially high-frequency
signals widespread in graphs with heterophily. Motivated by this limitation, we
propose a multi-view perspective and the usage of diverse pretext tasks to
capture different signals in graphs into embeddings. A novel framework, denoted
as Multi-view Graph Encoder(MVGE), is proposed, and a set of key designs are
identified. More specifically, a set of new pretext tasks are designed to
encode different types of signals, and a straightforward operation is
propxwosed to maintain both the commodity and personalization in both the
attribute and the structural levels. Extensive experiments on synthetic and
real-world network datasets show that the node representations learned with
MVGE achieve significant performance improvements in three different downstream
tasks, especially on graphs with heterophily. Source code is available at
\url{https://github.com/G-AILab/MVGE}.
Related papers
- Exploring Task Unification in Graph Representation Learning via Generative Approach [12.983429541410617]
Graphs are ubiquitous in real-world scenarios and encompass a diverse range of tasks, from node-, edge-, and graph-level tasks to transfer learning.
Recent endeavors aim to design a unified framework capable of generalizing across multiple graph tasks.
Among these, graph autoencoders (GAEs) have demonstrated their potential in effectively addressing various graph tasks.
We propose GA2E, a unified adversarially masked autoencoder capable of addressing the above challenges seamlessly.
arXiv Detail & Related papers (2024-03-21T12:14:02Z) - ALEX: Towards Effective Graph Transfer Learning with Noisy Labels [11.115297917940829]
We introduce a novel technique termed Balance Alignment and Information-aware Examination (ALEX) to address the problem of graph transfer learning.
ALEX first employs singular value decomposition to generate different views with crucial structural semantics, which help provide robust node representations.
Building on this foundation, an adversarial domain discriminator is incorporated for the implicit domain alignment of complex multi-modal distributions.
arXiv Detail & Related papers (2023-09-26T04:59:49Z) - SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - Group Contrastive Self-Supervised Learning on Graphs [101.45974132613293]
We study self-supervised learning on graphs using contrastive methods.
We argue that contrasting graphs in multiple subspaces enables graph encoders to capture more abundant characteristics.
arXiv Detail & Related papers (2021-07-20T22:09:21Z) - Multi-Level Graph Contrastive Learning [38.022118893733804]
We propose a Multi-Level Graph Contrastive Learning (MLGCL) framework for learning robust representation of graph data by contrasting space views of graphs.
The original graph is first-order approximation structure and contains uncertainty or error, while the $k$NN graph generated by encoding features preserves high-order proximity.
Extensive experiments indicate MLGCL achieves promising results compared with the existing state-of-the-art graph representation learning methods on seven datasets.
arXiv Detail & Related papers (2021-07-06T14:24:43Z) - A Robust and Generalized Framework for Adversarial Graph Embedding [73.37228022428663]
We propose a robust framework for adversarial graph embedding, named AGE.
AGE generates the fake neighbor nodes as the enhanced negative samples from the implicit distribution.
Based on this framework, we propose three models to handle three types of graph data.
arXiv Detail & Related papers (2021-05-22T07:05:48Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z) - Contrastive and Generative Graph Convolutional Networks for Graph-based
Semi-Supervised Learning [64.98816284854067]
Graph-based Semi-Supervised Learning (SSL) aims to transfer the labels of a handful of labeled data to the remaining massive unlabeled data via a graph.
A novel GCN-based SSL algorithm is presented in this paper to enrich the supervision signals by utilizing both data similarities and graph structure.
arXiv Detail & Related papers (2020-09-15T13:59:28Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.