QT-Routenet: Improved GNN generalization to larger 5G networks by
fine-tuning predictions from queueing theory
- URL: http://arxiv.org/abs/2207.06336v1
- Date: Wed, 13 Jul 2022 16:49:37 GMT
- Title: QT-Routenet: Improved GNN generalization to larger 5G networks by
fine-tuning predictions from queueing theory
- Authors: Bruno Klaus de Aquino Afonso, Lilian Berton
- Abstract summary: We tackle the problem of generalization when applying a model to a 5G network.
We propose to first extract robust features related to Queueing Theory (QT)
We then fine-tune the analytical baseline prediction using a modification of the Routenet Graph Neural Network (GNN) model.
- Score: 2.4366811507669124
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In order to promote the use of machine learning in 5G, the International
Telecommunication Union (ITU) proposed in 2021 the second edition of the ITU
AI/ML in 5G challenge, with over 1600 participants from 82 countries. This work
details the second place solution overall, which is also the winning solution
of the Graph Neural Networking Challenge 2021. We tackle the problem of
generalization when applying a model to a 5G network that may have longer paths
and larger link capacities than the ones observed in training. To achieve this,
we propose to first extract robust features related to Queueing Theory (QT),
and then fine-tune the analytical baseline prediction using a modification of
the Routenet Graph Neural Network (GNN) model. The proposed solution
generalizes much better than simply using Routenet, and manages to reduce the
analytical baseline's 10.42 mean absolute percent error to 1.45 (1.27 with an
ensemble). This suggests that making small changes to an approximate model that
is known to be robust can be an effective way to improve accuracy without
compromising generalization.
Related papers
- Unlocking the Theory Behind Scaling 1-Bit Neural Networks [7.578147116161996]
1-bit Large Language Models (LLMs) have emerged, showcasing an impressive combination of efficiency and performance that rivals traditional LLMs.
We present the first theoretical result that rigorously establishes a Scaling Law for 1-bit Neural Networks.
Our findings underscore the promising potential of scaling 1-bit neural networks, suggesting that int1 could become the standard in future neural network precision.
arXiv Detail & Related papers (2024-11-03T19:18:57Z) - Learning Load Balancing with GNN in MPTCP-Enabled Heterogeneous Networks [13.178956651532213]
We propose a graph neural network (GNN)-based model to tackle the LB problem for MP TCP-enabled HetNets.
Compared to the conventional deep neural network (DNN), the proposed GNN-based model exhibits two key strengths.
arXiv Detail & Related papers (2024-10-22T15:49:53Z) - Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch [72.26822499434446]
Auto-Train-Once (ATO) is an innovative network pruning algorithm designed to automatically reduce the computational and storage costs of DNNs.
We provide a comprehensive convergence analysis as well as extensive experiments, and the results show that our approach achieves state-of-the-art performance across various model architectures.
arXiv Detail & Related papers (2024-03-21T02:33:37Z) - Learning to Reweight for Graph Neural Network [63.978102332612906]
Graph Neural Networks (GNNs) show promising results for graph tasks.
Existing GNNs' generalization ability will degrade when there exist distribution shifts between testing and training graph data.
We propose a novel nonlinear graph decorrelation method, which can substantially improve the out-of-distribution generalization ability.
arXiv Detail & Related papers (2023-12-19T12:25:10Z) - Solving Large-scale Spatial Problems with Convolutional Neural Networks [88.31876586547848]
We employ transfer learning to improve training efficiency for large-scale spatial problems.
We propose that a convolutional neural network (CNN) can be trained on small windows of signals, but evaluated on arbitrarily large signals with little to no performance degradation.
arXiv Detail & Related papers (2023-06-14T01:24:42Z) - A New Perspective for Understanding Generalization Gap of Deep Neural
Networks Trained with Large Batch Sizes [14.822603738271138]
Deep neural networks (DNNs) are typically optimized using various forms of mini-batch gradient descent algorithm.
Many works report the progressive loss of model generalization when the training batch size is increased beyond some limits.
This is a scenario commonly referred to as generalization gap.
Our analysis suggests that large training batch size results in increased near-rank loss of units' activation.
arXiv Detail & Related papers (2022-10-21T18:23:12Z) - Elastic-Link for Binarized Neural Network [9.83865304744923]
"Elastic-Link" (EL) module enrich information flow within a BNN by adaptively adding real-valued input features to the subsequent convolutional output features.
EL produces a significant improvement on the challenging large-scale ImageNet dataset.
With the integration of ReActNet, it yields a new state-of-the-art result of 71.9% top-1 accuracy.
arXiv Detail & Related papers (2021-12-19T13:49:29Z) - On Topology Optimization and Routing in Integrated Access and Backhaul
Networks: A Genetic Algorithm-based Approach [70.85399600288737]
We study the problem of topology optimization and routing in IAB networks.
We develop efficient genetic algorithm-based schemes for both IAB node placement and non-IAB backhaul link distribution.
We discuss the main challenges for enabling mesh-based IAB networks.
arXiv Detail & Related papers (2021-02-14T21:52:05Z) - Graph Neural Networks for Scalable Radio Resource Management:
Architecture Design and Theoretical Analysis [31.372548374969387]
We propose to apply graph neural networks (GNNs) to solve large-scale radio resource management problems.
The proposed method is highly scalable and can solve the beamforming problem in an interference channel with $1000$ transceiver pairs within $6$ milliseconds on a single GPU.
arXiv Detail & Related papers (2020-07-15T11:43:32Z) - Fast Learning of Graph Neural Networks with Guaranteed Generalizability:
One-hidden-layer Case [93.37576644429578]
Graph neural networks (GNNs) have made great progress recently on learning from graph-structured data in practice.
We provide a theoretically-grounded generalizability analysis of GNNs with one hidden layer for both regression and binary classification problems.
arXiv Detail & Related papers (2020-06-25T00:45:52Z) - Optimization and Generalization Analysis of Transduction through
Gradient Boosting and Application to Multi-scale Graph Neural Networks [60.22494363676747]
It is known that the current graph neural networks (GNNs) are difficult to make themselves deep due to the problem known as over-smoothing.
Multi-scale GNNs are a promising approach for mitigating the over-smoothing problem.
We derive the optimization and generalization guarantees of transductive learning algorithms that include multi-scale GNNs.
arXiv Detail & Related papers (2020-06-15T17:06:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.