CGMN: A Contrastive Graph Matching Network for Self-Supervised Graph
Similarity Learning
- URL: http://arxiv.org/abs/2205.15083v1
- Date: Mon, 30 May 2022 13:20:26 GMT
- Title: CGMN: A Contrastive Graph Matching Network for Self-Supervised Graph
Similarity Learning
- Authors: Di Jin, Luzhi Wang, Yizhen Zheng, Xiang Li, Fei Jiang, Wei Lin, Shirui
Pan
- Abstract summary: We propose a contrastive graph matching network (CGMN) for self-supervised graph similarity learning.
We employ two strategies, namely cross-view interaction and cross-graph interaction, for effective node representation learning.
We transform node representations into graph-level representations via pooling operations for graph similarity computation.
- Score: 65.1042892570989
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Graph similarity learning refers to calculating the similarity score between
two graphs, which is required in many realistic applications, such as visual
tracking, graph classification, and collaborative filtering. As most of the
existing graph neural networks yield effective graph representations of a
single graph, little effort has been made for jointly learning two graph
representations and calculating their similarity score. In addition, existing
unsupervised graph similarity learning methods are mainly clustering-based,
which ignores the valuable information embodied in graph pairs. To this end, we
propose a contrastive graph matching network (CGMN) for self-supervised graph
similarity learning in order to calculate the similarity between any two input
graph objects. Specifically, we generate two augmented views for each graph in
a pair respectively. Then, we employ two strategies, namely cross-view
interaction and cross-graph interaction, for effective node representation
learning. The former is resorted to strengthen the consistency of node
representations in two views. The latter is utilized to identify node
differences between different graphs. Finally, we transform node
representations into graph-level representations via pooling operations for
graph similarity computation. We have evaluated CGMN on eight real-world
datasets, and the experiment results show that the proposed new approach is
superior to the state-of-the-art methods in graph similarity learning
downstream tasks.
Related papers
- Graph Self-supervised Learning with Accurate Discrepancy Learning [64.69095775258164]
We propose a framework that aims to learn the exact discrepancy between the original and the perturbed graphs, coined as Discrepancy-based Self-supervised LeArning (D-SLA)
We validate our method on various graph-related downstream tasks, including molecular property prediction, protein function prediction, and link prediction tasks, on which our model largely outperforms relevant baselines.
arXiv Detail & Related papers (2022-02-07T08:04:59Z) - Edge but not Least: Cross-View Graph Pooling [76.71497833616024]
This paper presents a cross-view graph pooling (Co-Pooling) method to better exploit crucial graph structure information.
Through cross-view interaction, edge-view pooling and node-view pooling seamlessly reinforce each other to learn more informative graph-level representations.
arXiv Detail & Related papers (2021-09-24T08:01:23Z) - Joint Graph Learning and Matching for Semantic Feature Correspondence [69.71998282148762]
We propose a joint emphgraph learning and matching network, named GLAM, to explore reliable graph structures for boosting graph matching.
The proposed method is evaluated on three popular visual matching benchmarks (Pascal VOC, Willow Object and SPair-71k)
It outperforms previous state-of-the-art graph matching methods by significant margins on all benchmarks.
arXiv Detail & Related papers (2021-09-01T08:24:02Z) - Multi-Level Graph Contrastive Learning [38.022118893733804]
We propose a Multi-Level Graph Contrastive Learning (MLGCL) framework for learning robust representation of graph data by contrasting space views of graphs.
The original graph is first-order approximation structure and contains uncertainty or error, while the $k$NN graph generated by encoding features preserves high-order proximity.
Extensive experiments indicate MLGCL achieves promising results compared with the existing state-of-the-art graph representation learning methods on seven datasets.
arXiv Detail & Related papers (2021-07-06T14:24:43Z) - Multilevel Graph Matching Networks for Deep Graph Similarity Learning [79.3213351477689]
We propose a multi-level graph matching network (MGMN) framework for computing the graph similarity between any pair of graph-structured objects.
To compensate for the lack of standard benchmark datasets, we have created and collected a set of datasets for both the graph-graph classification and graph-graph regression tasks.
Comprehensive experiments demonstrate that MGMN consistently outperforms state-of-the-art baseline models on both the graph-graph classification and graph-graph regression tasks.
arXiv Detail & Related papers (2020-07-08T19:48:19Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - Graph Partitioning and Graph Neural Network based Hierarchical Graph
Matching for Graph Similarity Computation [5.710312846460821]
Graph similarity aims to predict a similarity score between one pair of graphs to facilitate downstream applications.
We propose a graph partitioning and graph neural network-based model, called PSimGNN, to effectively resolve this issue.
PSimGNN outperforms state-of-the-art methods in graph similarity computation tasks using approximate Graph Edit Distance (GED) as the graph similarity metric.
arXiv Detail & Related papers (2020-05-16T15:01:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.