Learning Sub-Second Routing Optimization in Computer Networks requires Packet-Level Dynamics
- URL: http://arxiv.org/abs/2410.10377v1
- Date: Mon, 14 Oct 2024 11:03:46 GMT
- Title: Learning Sub-Second Routing Optimization in Computer Networks requires Packet-Level Dynamics
- Authors: Andreas Boltres, Niklas Freymuth, Patrick Jahnke, Holger Karl, Gerhard Neumann,
- Abstract summary: Reinforcement Learning can help to learn network representations that provide routing decisions.
We present $textitPackeRL$, the first packet-level Reinforcement Learning environment for routing in generic network topologies.
We also introduce two new algorithms for learning sub-second Routing Optimization.
- Score: 15.018408728324887
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Finding efficient routes for data packets is an essential task in computer networking. The optimal routes depend greatly on the current network topology, state and traffic demand, and they can change within milliseconds. Reinforcement Learning can help to learn network representations that provide routing decisions for possibly novel situations. So far, this has commonly been done using fluid network models. We investigate their suitability for millisecond-scale adaptations with a range of traffic mixes and find that packet-level network models are necessary to capture true dynamics, in particular in the presence of TCP traffic. To this end, we present $\textit{PackeRL}$, the first packet-level Reinforcement Learning environment for routing in generic network topologies. Our experiments confirm that learning-based strategies that have been trained in fluid environments do not generalize well to this more realistic, but more challenging setup. Hence, we also introduce two new algorithms for learning sub-second Routing Optimization. We present $\textit{M-Slim}$, a dynamic shortest-path algorithm that excels at high traffic volumes but is computationally hard to scale to large network topologies, and $\textit{FieldLines}$, a novel next-hop policy design that re-optimizes routing for any network topology within milliseconds without requiring any re-training. Both algorithms outperform current learning-based approaches as well as commonly used static baseline protocols in scenarios with high-traffic volumes. All findings are backed by extensive experiments in realistic network conditions in our fast and versatile training and evaluation framework.
Related papers
- Intelligent Routing Algorithm over SDN: Reusable Reinforcement Learning Approach [1.799933345199395]
We develop a reusable RL-aware, reusable routing algorithm, RLSR-Routing over SDN.
Our algorithm shows better performance in terms of load balancing than the traditional approaches.
It also has faster convergence than the non-reusable RL approach when finding paths for multiple traffic demands.
arXiv Detail & Related papers (2024-09-23T17:15:24Z) - Principled Architecture-aware Scaling of Hyperparameters [69.98414153320894]
Training a high-quality deep neural network requires choosing suitable hyperparameters, which is a non-trivial and expensive process.
In this work, we precisely characterize the dependence of initializations and maximal learning rates on the network architecture.
We demonstrate that network rankings can be easily changed by better training networks in benchmarks.
arXiv Detail & Related papers (2024-02-27T11:52:49Z) - A Deep Reinforcement Learning Approach for Adaptive Traffic Routing in
Next-gen Networks [1.1586742546971471]
Next-gen networks require automation and adaptively adjust network configuration based on traffic dynamics.
Traditional techniques that decide traffic policies are usually based on hand-crafted programming optimization and algorithms.
We develop a deep reinforcement learning (DRL) approach for adaptive traffic routing.
arXiv Detail & Related papers (2024-02-07T01:48:29Z) - XRoute Environment: A Novel Reinforcement Learning Environment for
Routing [8.797544401458476]
We introduce the XRoute Environment, a new reinforcement learning environment.
Agents are trained to select and route nets in an advanced, end-to-end routing framework.
The resulting environment is challenging, easy to use, customize and add additional scenarios.
arXiv Detail & Related papers (2023-05-23T08:46:25Z) - MARLIN: Soft Actor-Critic based Reinforcement Learning for Congestion
Control in Real Networks [63.24965775030673]
We propose a novel Reinforcement Learning (RL) approach to design generic Congestion Control (CC) algorithms.
Our solution, MARLIN, uses the Soft Actor-Critic algorithm to maximize both entropy and return.
We trained MARLIN on a real network with varying background traffic patterns to overcome the sim-to-real mismatch.
arXiv Detail & Related papers (2023-02-02T18:27:20Z) - ENERO: Efficient Real-Time Routing Optimization [2.830334160074889]
Traffic Engineering (TE) solutions must be able to achieve high performance real-time network operation.
Current TE technologies rely on hand-crafteds or computationally expensive solvers.
We propose Enero, an efficient real-time TE engine.
arXiv Detail & Related papers (2021-09-22T17:53:30Z) - Efficient Transfer Learning via Joint Adaptation of Network Architecture
and Weight [66.8543732597723]
Recent worksin neural architecture search (NAS) can aid transfer learning by establishing sufficient network search space.
We propose a novel framework consisting of two modules, the neural architecturesearch module for architecture transfer and the neural weight search module for weight transfer.
These two modules conduct search on thetarget task based on a reduced super-networks, so we only need to trainonce on the source task.
arXiv Detail & Related papers (2021-05-19T08:58:04Z) - Better than the Best: Gradient-based Improper Reinforcement Learning for
Network Scheduling [60.48359567964899]
We consider the problem of scheduling in constrained queueing networks with a view to minimizing packet delay.
We use a policy gradient based reinforcement learning algorithm that produces a scheduler that performs better than the available atomic policies.
arXiv Detail & Related papers (2021-05-01T10:18:34Z) - Relational Deep Reinforcement Learning for Routing in Wireless Networks [2.997420836766863]
We develop a distributed routing strategy based on deep reinforcement learning that generalizes to diverse traffic patterns, congestion levels, network connectivity, and link dynamics.
Our algorithm outperforms shortest path and backpressure routing with respect to packets delivered and delay per packet.
arXiv Detail & Related papers (2020-12-31T16:28:21Z) - Subset Sampling For Progressive Neural Network Learning [106.12874293597754]
Progressive Neural Network Learning is a class of algorithms that incrementally construct the network's topology and optimize its parameters based on the training data.
We propose to speed up this process by exploiting subsets of training data at each incremental training step.
Experimental results in object, scene and face recognition problems demonstrate that the proposed approach speeds up the optimization procedure considerably.
arXiv Detail & Related papers (2020-02-17T18:57:33Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.