Solve routing problems with a residual edge-graph attention neural
network
- URL: http://arxiv.org/abs/2105.02730v1
- Date: Thu, 6 May 2021 14:47:47 GMT
- Title: Solve routing problems with a residual edge-graph attention neural
network
- Authors: Kun Lei, Peng Guo, Yi Wang, Xiao Wu, Wenchao Zhao
- Abstract summary: In this paper, an end-to-end deep reinforcement learning framework is proposed to solve NP-hard optimization problems.
The proposed framework is aiming to improve the models in literacy in terms of the neural network model and the training algorithm.
Specifically, the average optimality gap is reduced from 4.53% (reported best citeR22) to 3.67% for TSP with 100 nodes and from 7.34% (reported best citeR22) to 6.68% for CVRP with 100 nodes when using the greedy decoding strategy.
- Score: 10.42872026679616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For NP-hard combinatorial optimization problems, it is usually difficult to
find high-quality solutions in polynomial time. The design of either an exact
algorithm or an approximate algorithm for these problems often requires
significantly specialized knowledge. Recently, deep learning methods provide
new directions to solve such problems. In this paper, an end-to-end deep
reinforcement learning framework is proposed to solve this type of
combinatorial optimization problems. This framework can be applied to different
problems with only slight changes of input (for example, for a traveling
salesman problem (TSP), the input is the two-dimensional coordinates of nodes;
while for a capacity-constrained vehicle routing problem (CVRP), the input is
simply changed to three-dimensional vectors including the two-dimensional
coordinates and the customer demands of nodes), masks and decoder context
vectors. The proposed framework is aiming to improve the models in literacy in
terms of the neural network model and the training algorithm. The solution
quality of TSP and the CVRP up to 100 nodes are significantly improved via our
framework. Specifically, the average optimality gap is reduced from 4.53\%
(reported best \cite{R22}) to 3.67\% for TSP with 100 nodes and from 7.34\%
(reported best \cite{R22}) to 6.68\% for CVRP with 100 nodes when using the
greedy decoding strategy. Furthermore, our framework uses about 1/3$\sim$3/4
training samples compared with other existing learning methods while achieving
better results. The results performed on randomly generated instances and the
benchmark instances from TSPLIB and CVRPLIB confirm that our framework has a
linear running time on the problem size (number of nodes) during the testing
phase, and has a good generalization performance from random instance training
to real-world instance testing.
Related papers
- Symmetry-preserving graph attention network to solve routing problems at
multiple resolutions [1.9304772860080408]
We introduce the first-ever completely equivariant model and training to solve problems.
It is essential to capture the multiscale structure of the input graph.
We propose a Multiresolution scheme in combination with Equi Graph Attention network (mEGAT) architecture.
arXiv Detail & Related papers (2023-10-24T06:22:20Z) - On Statistical Learning of Branch and Bound for Vehicle Routing
Optimization [3.6922704509753084]
We train neural networks to emulate the decision-making process of the computationally expensive Strong Branching strategy.
We find that this approach can match or improve upon the performance of the branch and bound algorithm.
arXiv Detail & Related papers (2023-10-15T23:59:57Z) - Revisit the Algorithm Selection Problem for TSP with Spatial Information
Enhanced Graph Neural Networks [4.084365114504618]
This paper revisits the algorithm selection problem for Euclidean Traveling Salesman Problem (TSP)
We propose a novel Graph Neural Network (GNN), called GINES.
GINES takes the coordinates of cities and distances between cities as input.
It is better than the traditional handcrafted feature-based approach on one dataset.
arXiv Detail & Related papers (2023-02-08T13:14:20Z) - Towards Better Out-of-Distribution Generalization of Neural Algorithmic
Reasoning Tasks [51.8723187709964]
We study the OOD generalization of neural algorithmic reasoning tasks.
The goal is to learn an algorithm from input-output pairs using deep neural networks.
arXiv Detail & Related papers (2022-11-01T18:33:20Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - TCT: Convexifying Federated Learning using Bootstrapped Neural Tangent
Kernels [141.29156234353133]
State-of-the-art convex learning methods can perform far worse than their centralized counterparts when clients have dissimilar data distributions.
We show this disparity can largely be attributed to challenges presented by non-NISTity.
We propose a Train-Convexify neural network (TCT) procedure to sidestep this issue.
arXiv Detail & Related papers (2022-07-13T16:58:22Z) - A Continuous Optimisation Benchmark Suite from Neural Network Regression [0.0]
Training neural networks is an optimisation task that has gained prominence with the recent successes of deep learning.
gradient descent variants are by far the most common choice with their trusted good performance on large-scale machine learning tasks.
We contribute CORNN, a suite for benchmarking the performance of any continuous black-box algorithm on neural network training problems.
arXiv Detail & Related papers (2021-09-12T20:24:11Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Channel Assignment in Uplink Wireless Communication using Machine
Learning Approach [54.012791474906514]
This letter investigates a channel assignment problem in uplink wireless communication systems.
Our goal is to maximize the sum rate of all users subject to integer channel assignment constraints.
Due to high computational complexity, machine learning approaches are employed to obtain computational efficient solutions.
arXiv Detail & Related papers (2020-01-12T15:54:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.