Learning Enhanced Optimisation for Routing Problems
- URL: http://arxiv.org/abs/2109.08345v1
- Date: Fri, 17 Sep 2021 04:47:26 GMT
- Title: Learning Enhanced Optimisation for Routing Problems
- Authors: Nasrin Sultana, Jeffrey Chan, Tabinda Sarwar, Babak Abbasi, A. K. Qin
- Abstract summary: "Learning to Guide Local Search" (L2GLS) is a learning-based approach for routing problems.
L2GLS combines local search (LS) operators' strengths with penalty terms to escape local optimals.
We show that L2GLS achieves the new state-of-the-art results on larger TSP and CVRP over other machine learning methods.
- Score: 3.747361228408185
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep learning approaches have shown promising results in solving routing
problems. However, there is still a substantial gap in solution quality between
machine learning and operations research algorithms. Recently, another line of
research has been introduced that fuses the strengths of machine learning and
operational research algorithms. In particular, search perturbation operators
have been used to improve the solution. Nevertheless, using the perturbation
may not guarantee a quality solution. This paper presents "Learning to Guide
Local Search" (L2GLS), a learning-based approach for routing problems that uses
a penalty term and reinforcement learning to adaptively adjust search efforts.
L2GLS combines local search (LS) operators' strengths with penalty terms to
escape local optimals. Routing problems have many practical applications, often
presetting larger instances that are still challenging for many existing
algorithms introduced in the learning to optimise field. We show that L2GLS
achieves the new state-of-the-art results on larger TSP and CVRP over other
machine learning methods.
Related papers
- Decision Transformer for Enhancing Neural Local Search on the Job Shop Scheduling Problem [10.316443594063173]
Job shop scheduling problem (JSSP) and its solution algorithms have been of enduring interest in both academia and industry for decades.
In recent years, machine learning (ML) is playing an increasingly important role in advancing existing and building new solutions for the JSSP, aiming to find better solutions in shorter times.
We build on top of a state-of-the-art deep reinforcement learning (DRL) agent, called Neural Local Search (NLS), which can efficiently and effectively control a large local neighborhood search on the JSSP.
arXiv Detail & Related papers (2024-09-04T13:33:38Z) - New Characterizations and Efficient Local Search for General Integer
Linear Programming [17.80124240223163]
This work proposes new characterizations of linear programming (ILP) with the concept of boundary solutions.
We develop a new local search algorithm Local-ILP, which is efficient for solving general ILP validated on a large heterogeneous problem dataset.
Experiments conducted on the MIPLIB dataset show the effectiveness of our algorithm in solving large-scale hard ILP problems.
arXiv Detail & Related papers (2023-04-29T07:22:07Z) - MARLIN: Soft Actor-Critic based Reinforcement Learning for Congestion
Control in Real Networks [63.24965775030673]
We propose a novel Reinforcement Learning (RL) approach to design generic Congestion Control (CC) algorithms.
Our solution, MARLIN, uses the Soft Actor-Critic algorithm to maximize both entropy and return.
We trained MARLIN on a real network with varying background traffic patterns to overcome the sim-to-real mismatch.
arXiv Detail & Related papers (2023-02-02T18:27:20Z) - Sample-Efficient, Exploration-Based Policy Optimisation for Routing
Problems [2.6782615615913348]
This paper presents a new reinforcement learning approach that is based on entropy.
In addition, we design an off-policy-based reinforcement learning technique that maximises the expected return.
We show that our model can generalise to various route problems.
arXiv Detail & Related papers (2022-05-31T09:51:48Z) - Reinforcement Learning for Branch-and-Bound Optimisation using
Retrospective Trajectories [72.15369769265398]
Machine learning has emerged as a promising paradigm for branching.
We propose retro branching; a simple yet effective approach to RL for branching.
We outperform the current state-of-the-art RL branching algorithm by 3-5x and come within 20% of the best IL method's performance on MILPs with 500 constraints and 1000 variables.
arXiv Detail & Related papers (2022-05-28T06:08:07Z) - Accelerated Policy Learning with Parallel Differentiable Simulation [59.665651562534755]
We present a differentiable simulator and a new policy learning algorithm (SHAC)
Our algorithm alleviates problems with local minima through a smooth critic function.
We show substantial improvements in sample efficiency and wall-clock time over state-of-the-art RL and differentiable simulation-based algorithms.
arXiv Detail & Related papers (2022-04-14T17:46:26Z) - Contextual Exploration Using a Linear Approximation Method Based on
Satisficing [0.0]
The amount of exploration required for learning is often quite large.
Deep reinforcement learning also has super-human performance in that no human being would be able to achieve such amounts of exploration.
We propose Linear RS (LinRS), which is a type of satisficing algorithm and a linear extension of risk-sensitive satisficing (RS)
arXiv Detail & Related papers (2021-12-13T07:14:01Z) - An Overview and Experimental Study of Learning-based Optimization
Algorithms for Vehicle Routing Problem [49.04543375851723]
Vehicle routing problem (VRP) is a typical discrete optimization problem.
Many studies consider learning-based optimization algorithms to solve VRP.
This paper reviews recent advances in this field and divides relevant approaches into end-to-end approaches and step-by-step approaches.
arXiv Detail & Related papers (2021-07-15T02:13:03Z) - A Two-stage Framework and Reinforcement Learning-based Optimization
Algorithms for Complex Scheduling Problems [54.61091936472494]
We develop a two-stage framework, in which reinforcement learning (RL) and traditional operations research (OR) algorithms are combined together.
The scheduling problem is solved in two stages, including a finite Markov decision process (MDP) and a mixed-integer programming process, respectively.
Results show that the proposed algorithms could stably and efficiently obtain satisfactory scheduling schemes for agile Earth observation satellite scheduling problems.
arXiv Detail & Related papers (2021-03-10T03:16:12Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - Learning 2-opt Heuristics for the Traveling Salesman Problem via Deep
Reinforcement Learning [2.4565068569913384]
We propose to learn a local search gradient based on 2-opt operators via deep reinforcement learning.
We show that the learned policies can improve even over random initial solutions and approach near-optimal solutions at a faster rate than previous state-of-the-art deep learning methods.
arXiv Detail & Related papers (2020-04-03T14:51:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.