Geometry Contrastive Learning on Heterogeneous Graphs
- URL: http://arxiv.org/abs/2206.12547v1
- Date: Sat, 25 Jun 2022 03:54:53 GMT
- Title: Geometry Contrastive Learning on Heterogeneous Graphs
- Authors: Shichao Zhu, Chuan Zhou, Anfeng Cheng, Shirui Pan, Shuaiqiang Wang,
Dawei Yin, Bin Wang
- Abstract summary: This paper proposes a novel self-supervised learning method, termed as Geometry Contrastive Learning (GCL)
GCL views a heterogeneous graph from Euclidean and hyperbolic perspective simultaneously, aiming to make a strong merger of the ability of modeling rich semantics and complex structures.
Extensive experiments on four benchmarks data sets show that the proposed approach outperforms the strong baselines.
- Score: 50.58523799455101
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning (especially contrastive learning) methods on
heterogeneous graphs can effectively get rid of the dependence on supervisory
data. Meanwhile, most existing representation learning methods embed the
heterogeneous graphs into a single geometric space, either Euclidean or
hyperbolic. This kind of single geometric view is usually not enough to observe
the complete picture of heterogeneous graphs due to their rich semantics and
complex structures. Under these observations, this paper proposes a novel
self-supervised learning method, termed as Geometry Contrastive Learning (GCL),
to better represent the heterogeneous graphs when supervisory data is
unavailable. GCL views a heterogeneous graph from Euclidean and hyperbolic
perspective simultaneously, aiming to make a strong merger of the ability of
modeling rich semantics and complex structures, which is expected to bring in
more benefits for downstream tasks. GCL maximizes the mutual information
between two geometric views by contrasting representations at both local-local
and local-global semantic levels. Extensive experiments on four benchmarks data
sets show that the proposed approach outperforms the strong baselines,
including both unsupervised methods and supervised methods, on three tasks,
including node classification, node clustering and similarity search.
Related papers
- When Heterophily Meets Heterogeneous Graphs: Latent Graphs Guided Unsupervised Representation Learning [6.2167203720326025]
Unsupervised heterogeneous graph representation learning (UHGRL) has gained increasing attention due to its significance in handling practical graphs without labels.
We define semantic heterophily and propose an innovative framework called Latent Graphs Guided Unsupervised Representation Learning (LatGRL) to handle this problem.
arXiv Detail & Related papers (2024-09-01T10:25:06Z) - Localized Contrastive Learning on Graphs [110.54606263711385]
We introduce a simple yet effective contrastive model named Localized Graph Contrastive Learning (Local-GCL)
In spite of its simplicity, Local-GCL achieves quite competitive performance in self-supervised node representation learning tasks on graphs with various scales and properties.
arXiv Detail & Related papers (2022-12-08T23:36:00Z) - Revisiting Heterophily in Graph Convolution Networks by Learning
Representations Across Topological and Feature Spaces [20.775165967590173]
Graph convolution networks (GCNs) have been enormously successful in learning representations over several graph-based machine learning tasks.
We argue that by learning graph representations across two spaces i.e., topology and feature space GCNs can address heterophily.
We experimentally demonstrate the performance of the proposed GCN framework over semi-supervised node classification task.
arXiv Detail & Related papers (2022-11-01T16:21:10Z) - Heterogeneous Graph Neural Networks using Self-supervised Reciprocally
Contrastive Learning [102.9138736545956]
Heterogeneous graph neural network (HGNN) is a very popular technique for the modeling and analysis of heterogeneous graphs.
We develop for the first time a novel and robust heterogeneous graph contrastive learning approach, namely HGCL, which introduces two views on respective guidance of node attributes and graph topologies.
In this new approach, we adopt distinct but most suitable attribute and topology fusion mechanisms in the two views, which are conducive to mining relevant information in attributes and topologies separately.
arXiv Detail & Related papers (2022-04-30T12:57:02Z) - ACTIVE:Augmentation-Free Graph Contrastive Learning for Partial
Multi-View Clustering [52.491074276133325]
We propose an augmentation-free graph contrastive learning framework to solve the problem of partial multi-view clustering.
The proposed approach elevates instance-level contrastive learning and missing data inference to the cluster-level, effectively mitigating the impact of individual missing data on clustering.
arXiv Detail & Related papers (2022-03-01T02:32:25Z) - Effective and Efficient Graph Learning for Multi-view Clustering [173.8313827799077]
We propose an effective and efficient graph learning model for multi-view clustering.
Our method exploits the view-similar between graphs of different views by the minimization of tensor Schatten p-norm.
Our proposed algorithm is time-economical and obtains the stable results and scales well with the data size.
arXiv Detail & Related papers (2021-08-15T13:14:28Z) - Graph Geometry Interaction Learning [41.10468385822182]
We develop a novel Geometry Interaction Learning (GIL) method for graphs, a well-suited and efficient alternative for learning abundant geometric properties in graph.
Our method endows each node the freedom to determine the importance of each geometry space via a flexible dual feature interaction learning and probability assembling mechanism.
Promising experimental results are presented for five benchmark datasets on node classification and link prediction tasks.
arXiv Detail & Related papers (2020-10-23T02:40:28Z) - Structured Graph Learning for Clustering and Semi-supervised
Classification [74.35376212789132]
We propose a graph learning framework to preserve both the local and global structure of data.
Our method uses the self-expressiveness of samples to capture the global structure and adaptive neighbor approach to respect the local structure.
Our model is equivalent to a combination of kernel k-means and k-means methods under certain condition.
arXiv Detail & Related papers (2020-08-31T08:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.