IGNNITION: Bridging the Gap Between Graph Neural Networks and Networking
Systems
- URL: http://arxiv.org/abs/2109.06715v1
- Date: Tue, 14 Sep 2021 14:28:21 GMT
- Title: IGNNITION: Bridging the Gap Between Graph Neural Networks and Networking
Systems
- Authors: David Pujol-Perich, Jos\'e Su\'arez-Varela, Miquel Ferriol, Shihan
Xiao, Bo Wu, Albert Cabellos-Aparicio, Pere Barlet-Ros
- Abstract summary: We present IGNNITION, a novel open-source framework that enables fast prototyping of Graph Neural Networks (GNNs) for networking systems.
IGNNITION is based on an intuitive high-level abstraction that hides the complexity behind GNNs.
Our results show that the GNN models produced by IGNNITION are equivalent in terms of accuracy and performance to their native implementations.
- Score: 4.1591055164123665
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent years have seen the vast potential of Graph Neural Networks (GNN) in
many fields where data is structured as graphs (e.g., chemistry, recommender
systems). In particular, GNNs are becoming increasingly popular in the field of
networking, as graphs are intrinsically present at many levels (e.g., topology,
routing). The main novelty of GNNs is their ability to generalize to other
networks unseen during training, which is an essential feature for developing
practical Machine Learning (ML) solutions for networking. However, implementing
a functional GNN prototype is currently a cumbersome task that requires strong
skills in neural network programming. This poses an important barrier to
network engineers that often do not have the necessary ML expertise. In this
article, we present IGNNITION, a novel open-source framework that enables fast
prototyping of GNNs for networking systems. IGNNITION is based on an intuitive
high-level abstraction that hides the complexity behind GNNs, while still
offering great flexibility to build custom GNN architectures. To showcase the
versatility and performance of this framework, we implement two
state-of-the-art GNN models applied to different networking use cases. Our
results show that the GNN models produced by IGNNITION are equivalent in terms
of accuracy and performance to their native implementations in TensorFlow.
Related papers
- LinkSAGE: Optimizing Job Matching Using Graph Neural Networks [12.088731514483104]
We present LinkSAGE, an innovative framework that integrates Graph Neural Networks (GNNs) into large-scale personalized job matching systems.
Our approach capitalizes on a novel job marketplace graph, the largest and most intricate of its kind in industry, with billions of nodes and edges.
A key innovation in LinkSAGE is its training and serving methodology, which effectively combines inductive graph learning on a heterogeneous, evolving graph with an encoder-decoder GNN model.
arXiv Detail & Related papers (2024-02-20T23:49:25Z) - Distributed Graph Neural Network Training: A Survey [51.77035975191926]
Graph neural networks (GNNs) are a type of deep learning models that are trained on graphs and have been successfully applied in various domains.
Despite the effectiveness of GNNs, it is still challenging for GNNs to efficiently scale to large graphs.
As a remedy, distributed computing becomes a promising solution of training large-scale GNNs.
arXiv Detail & Related papers (2022-11-01T01:57:00Z) - Scaling Graph-based Deep Learning models to larger networks [2.946140899052065]
Graph Neural Networks (GNN) have shown a strong potential to be integrated into commercial products for network control and management.
This paper presents a GNN-based solution that can effectively scale to larger networks including higher link capacities and aggregated traffic on links.
arXiv Detail & Related papers (2021-10-04T09:04:19Z) - AdaGNN: A multi-modal latent representation meta-learner for GNNs based
on AdaBoosting [0.38073142980733]
Graph Neural Networks (GNNs) focus on extracting intrinsic network features.
We propose boosting-based meta learner for GNNs.
AdaGNN performs exceptionally well for applications with rich and diverse node neighborhood information.
arXiv Detail & Related papers (2021-08-14T03:07:26Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Learning to Execute Programs with Instruction Pointer Attention Graph
Neural Networks [55.98291376393561]
Graph neural networks (GNNs) have emerged as a powerful tool for learning software engineering tasks.
Recurrent neural networks (RNNs) are well-suited to long sequential chains of reasoning, but they do not naturally incorporate program structure.
We introduce a novel GNN architecture, the Instruction Pointer Attention Graph Neural Networks (IPA-GNN), which improves systematic generalization on the task of learning to execute programs.
arXiv Detail & Related papers (2020-10-23T19:12:30Z) - A Practical Tutorial on Graph Neural Networks [49.919443059032226]
Graph neural networks (GNNs) have recently grown in popularity in the field of artificial intelligence (AI)
This tutorial exposes the power and novelty of GNNs to AI practitioners.
arXiv Detail & Related papers (2020-10-11T12:36:17Z) - Computing Graph Neural Networks: A Survey from Algorithms to
Accelerators [2.491032752533246]
Graph Neural Networks (GNNs) have exploded onto the machine learning scene in recent years owing to their capability to model and learn from graph-structured data.
This paper aims to make two main contributions: a review of the field of GNNs is presented from the perspective of computing.
An in-depth analysis of current software and hardware acceleration schemes is provided.
arXiv Detail & Related papers (2020-09-30T22:29:27Z) - Architectural Implications of Graph Neural Networks [17.01480604968118]
Graph neural networks (GNN) represent an emerging line of deep learning models that operate on graph structures.
GNN is not as well understood in the system and architecture community as its counterparts such as multi-layer perceptrons and convolutional neural networks.
arXiv Detail & Related papers (2020-09-02T03:36:24Z) - Attentive Graph Neural Networks for Few-Shot Learning [74.01069516079379]
Graph Neural Networks (GNN) has demonstrated the superior performance in many challenging applications, including the few-shot learning tasks.
Despite its powerful capacity to learn and generalize the model from few samples, GNN usually suffers from severe over-fitting and over-smoothing as the model becomes deep.
We propose a novel Attentive GNN to tackle these challenges, by incorporating a triple-attention mechanism.
arXiv Detail & Related papers (2020-07-14T07:43:09Z) - GPT-GNN: Generative Pre-Training of Graph Neural Networks [93.35945182085948]
Graph neural networks (GNNs) have been demonstrated to be powerful in modeling graph-structured data.
We present the GPT-GNN framework to initialize GNNs by generative pre-training.
We show that GPT-GNN significantly outperforms state-of-the-art GNN models without pre-training by up to 9.1% across various downstream tasks.
arXiv Detail & Related papers (2020-06-27T20:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.