TIDE: Time Derivative Diffusion for Deep Learning on Graphs
- URL: http://arxiv.org/abs/2212.02483v3
- Date: Fri, 15 Sep 2023 14:45:39 GMT
- Title: TIDE: Time Derivative Diffusion for Deep Learning on Graphs
- Authors: Maysam Behmanesh, Maximilian Krahn, Maks Ovsjanikov
- Abstract summary: A prominent paradigm for graph neural networks is based on the message-passing framework.
In this framework, information communication is realized only between neighboring nodes.
We present a novel method based on time derivative graph diffusion (TIDE) to overcome these structural limitations.
- Score: 31.01454180524729
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A prominent paradigm for graph neural networks is based on the
message-passing framework. In this framework, information communication is
realized only between neighboring nodes. The challenge of approaches that use
this paradigm is to ensure efficient and accurate long-distance communication
between nodes, as deep convolutional networks are prone to oversmoothing. In
this paper, we present a novel method based on time derivative graph diffusion
(TIDE) to overcome these structural limitations of the message-passing
framework. Our approach allows for optimizing the spatial extent of diffusion
across various tasks and network channels, thus enabling medium and
long-distance communication efficiently. Furthermore, we show that our
architecture design also enables local message-passing and thus inherits from
the capabilities of local message-passing approaches. We show that on both
widely used graph benchmarks and synthetic mesh and graph datasets, the
proposed framework outperforms state-of-the-art methods by a significant margin
Related papers
- NetDiff: Deep Graph Denoising Diffusion for Ad Hoc Network Topology Generation [1.6768151308423371]
We introduce NetDiff, a graph denoising diffusion probabilistic architecture that generates wireless ad hoc network link topologies.
Our results show that the generated links are realistic, present structural properties similar to the dataset graphs', and require only minor corrections and verification steps to be operational.
arXiv Detail & Related papers (2024-10-09T15:39:49Z) - Task-Oriented Communication for Graph Data: A Graph Information Bottleneck Approach [12.451324619122405]
This paper introduces a method to extract a smaller, task-focused subgraph that maintains key information while reducing communication overhead.
Our approach utilizes graph neural networks (GNNs) and the graph information bottleneck (GIB) principle to create a compact, informative, and robust graph representation suitable for transmission.
arXiv Detail & Related papers (2024-09-04T14:01:56Z) - Network Alignment with Transferable Graph Autoencoders [79.89704126746204]
We propose a novel graph autoencoder architecture designed to extract powerful and robust node embeddings.
We prove that the generated embeddings are associated with the eigenvalues and eigenvectors of the graphs.
Our proposed framework also leverages transfer learning and data augmentation to achieve efficient network alignment at a very large scale without retraining.
arXiv Detail & Related papers (2023-10-05T02:58:29Z) - Learning How to Propagate Messages in Graph Neural Networks [55.2083896686782]
This paper studies the problem of learning message propagation strategies for graph neural networks (GNNs)
We introduce the optimal propagation steps as latent variables to help find the maximum-likelihood estimation of the GNN parameters.
Our proposed framework can effectively learn personalized and interpretable propagate strategies of messages in GNNs.
arXiv Detail & Related papers (2023-10-01T15:09:59Z) - Learning State-Augmented Policies for Information Routing in
Communication Networks [92.59624401684083]
We develop a novel State Augmentation (SA) strategy to maximize the aggregate information at source nodes using graph neural network (GNN) architectures.
We leverage an unsupervised learning procedure to convert the output of the GNN architecture to optimal information routing strategies.
In the experiments, we perform the evaluation on real-time network topologies to validate our algorithms.
arXiv Detail & Related papers (2023-09-30T04:34:25Z) - Building Shortcuts between Distant Nodes with Biaffine Mapping for Graph
Convolutional Networks [18.160610500658183]
We introduce Biaffine technique to improve the expressiveness of graph convolutional networks with a shallow architecture.
Our method is to learn direct dependency on long-distance neighbors for nodes, with which only one-hop message passing is capable of capturing rich information for node representation.
arXiv Detail & Related papers (2023-02-17T06:39:47Z) - DPGNN: Dual-Perception Graph Neural Network for Representation Learning [21.432960458513826]
Graph neural networks (GNNs) have drawn increasing attention in recent years and achieved remarkable performance in many graph-based tasks.
Most existing GNNs are based on the message-passing paradigm to iteratively aggregate neighborhood information in a single topology space.
We present a novel message-passing paradigm, based on the properties of multi-step message source, node-specific message output, and multi-space message interaction.
arXiv Detail & Related papers (2021-10-15T05:47:26Z) - Towards Efficient Scene Understanding via Squeeze Reasoning [71.1139549949694]
We propose a novel framework called Squeeze Reasoning.
Instead of propagating information on the spatial map, we first learn to squeeze the input feature into a channel-wise global vector.
We show that our approach can be modularized as an end-to-end trained block and can be easily plugged into existing networks.
arXiv Detail & Related papers (2020-11-06T12:17:01Z) - GCN for HIN via Implicit Utilization of Attention and Meta-paths [104.24467864133942]
Heterogeneous information network (HIN) embedding aims to map the structure and semantic information in a HIN to distributed representations.
We propose a novel neural network method via implicitly utilizing attention and meta-paths.
We first use the multi-layer graph convolutional network (GCN) framework, which performs a discriminative aggregation at each layer.
We then give an effective relaxation and improvement via introducing a new propagation operation which can be separated from aggregation.
arXiv Detail & Related papers (2020-07-06T11:09:40Z) - Constructing Geographic and Long-term Temporal Graph for Traffic
Forecasting [88.5550074808201]
We propose Geographic and Long term Temporal Graph Convolutional Recurrent Neural Network (GLT-GCRNN) for traffic forecasting.
In this work, we propose a novel framework for traffic forecasting that learns the rich interactions between roads sharing similar geographic or longterm temporal patterns.
arXiv Detail & Related papers (2020-04-23T03:50:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.