On a linear fused Gromov-Wasserstein distance for graph structured data
- URL: http://arxiv.org/abs/2203.04711v1
- Date: Wed, 9 Mar 2022 13:43:18 GMT
- Title: On a linear fused Gromov-Wasserstein distance for graph structured data
- Authors: Dai Hai Nguyen, Koji Tsuda
- Abstract summary: We propose a novel distance between two graphs, named linearFGW, defined as the Euclidean distance between their embeddings.
The advantages of the proposed distance are twofold: 1) it can take into account node feature and structure of graphs for measuring the similarity between graphs in a kernel-based framework, 2) it can be much faster for computing kernel matrix than pairwise OT-based distances.
- Score: 2.360534864805446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a framework for embedding graph structured data into a vector
space, taking into account node features and topology of a graph into the
optimal transport (OT) problem. Then we propose a novel distance between two
graphs, named linearFGW, defined as the Euclidean distance between their
embeddings. The advantages of the proposed distance are twofold: 1) it can take
into account node feature and structure of graphs for measuring the similarity
between graphs in a kernel-based framework, 2) it can be much faster for
computing kernel matrix than pairwise OT-based distances, particularly fused
Gromov-Wasserstein, making it possible to deal with large-scale data sets.
After discussing theoretical properties of linearFGW, we demonstrate
experimental results on classification and clustering tasks, showing the
effectiveness of the proposed linearFGW.
Related papers
- Exploiting Edge Features in Graphs with Fused Network Gromov-Wasserstein
Distance [18.522233517515975]
We introduce an extension of Gromov-Wasserstein distance for comparing graphs whose both nodes and edges have features.
We empirically show the effectiveness of the novel distance in learning tasks where graphs occur in either input space or output space.
arXiv Detail & Related papers (2023-09-28T17:05:03Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Graph Fourier MMD for Signals on Graphs [67.68356461123219]
We propose a novel distance between distributions and signals on graphs.
GFMMD is defined via an optimal witness function that is both smooth on the graph and maximizes difference in expectation.
We showcase it on graph benchmark datasets as well as on single cell RNA-sequencing data analysis.
arXiv Detail & Related papers (2023-06-05T00:01:17Z) - Efficient Graph Field Integrators Meet Point Clouds [59.27295475120132]
We present two new classes of algorithms for efficient field integration on graphs encoding point clouds.
The first class, SeparatorFactorization(SF), leverages the bounded genus of point cloud mesh graphs, while the second class, RFDiffusion(RFD), uses popular epsilon-nearest-neighbor graph representations for point clouds.
arXiv Detail & Related papers (2023-02-02T08:33:36Z) - Template based Graph Neural Network with Optimal Transport Distances [11.56532171513328]
Current Graph Neural Networks (GNN) architectures rely on two important components: node features embedding through message passing, and aggregation with a specialized form of pooling.
We propose in this work a novel point of view, which places distances to some learnable graph templates at the core of the graph representation.
This distance embedding is constructed thanks to an optimal transport distance: the Fused Gromov-Wasserstein (FGW) distance.
arXiv Detail & Related papers (2022-05-31T12:24:01Z) - Embedding Graphs on Grassmann Manifold [31.42901131602713]
This paper develops a new graph representation learning scheme, namely EGG, which embeds approximated second-order graph characteristics into a Grassmann manifold.
The effectiveness of EGG is demonstrated using both clustering and classification tasks at the node level and graph level.
arXiv Detail & Related papers (2022-05-30T12:56:24Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - Generalized Spectral Clustering via Gromov-Wasserstein Learning [10.558951653323286]
We establish a bridge between spectral clustering and Gromov-Wasserstein Learning (GWL)
GWL is a recent optimal transport-based approach to graph partitioning.
We show that when comparing against a two-node template graph using the heat kernel at the infinite time limit, the resulting partition agrees with the partition produced by the Fiedler vector.
arXiv Detail & Related papers (2020-06-07T14:29:32Z) - Wasserstein-based Graph Alignment [56.84964475441094]
We cast a new formulation for the one-to-many graph alignment problem, which aims at matching a node in the smaller graph with one or more nodes in the larger graph.
We show that our method leads to significant improvements with respect to the state-of-the-art algorithms for each of these tasks.
arXiv Detail & Related papers (2020-03-12T22:31:59Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.