Moco: A Learnable Meta Optimizer for Combinatorial Optimization
- URL: http://arxiv.org/abs/2402.04915v2
- Date: Fri, 9 Feb 2024 15:12:42 GMT
- Title: Moco: A Learnable Meta Optimizer for Combinatorial Optimization
- Authors: Tim Dernedde, Daniela Thyssens, S\"oren Dittrich, Maximilian
Stubbemann, Lars Schmidt-Thieme
- Abstract summary: Moco learns a graph neural network that updates the solution construction procedure based on features extracted from the current search state.
This meta training procedure targets the overall best solution found during the search procedure given information such as the search budget.
Moco is a fully learnable meta that does not utilize any problem specific local search or decomposition.
- Score: 5.359176539960004
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Relevant combinatorial optimization problems (COPs) are often NP-hard. While
they have been tackled mainly via handcrafted heuristics in the past, advances
in neural networks have motivated the development of general methods to learn
heuristics from data. Many approaches utilize a neural network to directly
construct a solution, but are limited in further improving based on already
constructed solutions at inference time. Our approach, Moco, learns a graph
neural network that updates the solution construction procedure based on
features extracted from the current search state. This meta training procedure
targets the overall best solution found during the search procedure given
information such as the search budget. This allows Moco to adapt to varying
circumstances such as different computational budgets. Moco is a fully
learnable meta optimizer that does not utilize any problem specific local
search or decomposition. We test Moco on the Traveling Salesman Problem (TSP)
and Maximum Independent Set (MIS) and show that it outperforms other approaches
on MIS and is overall competitive on the TSP, especially outperforming related
approaches, partially even if they use additional local search.
Related papers
- MARCO: A Memory-Augmented Reinforcement Framework for Combinatorial Optimization [44.24494442399324]
This paper introduces a versatile framework, referred to as Memory-Augmented Reinforcement for Combinatorial Optimization (MARCO)
MARCO stores data collected throughout the optimization trajectory and retrieves contextually relevant information at each state.
Thanks to the parallel nature of NCO models, several search threads can run simultaneously, all sharing the same memory module.
arXiv Detail & Related papers (2024-08-05T03:15:21Z) - Can Learned Optimization Make Reinforcement Learning Less Difficult? [70.5036361852812]
We consider whether learned optimization can help overcome reinforcement learning difficulties.
Our method, Learned Optimization for Plasticity, Exploration and Non-stationarity (OPEN), meta-learns an update rule whose input features and output structure are informed by previously proposed to these difficulties.
arXiv Detail & Related papers (2024-07-09T17:55:23Z) - Let the Flows Tell: Solving Graph Combinatorial Optimization Problems
with GFlowNets [86.43523688236077]
Combinatorial optimization (CO) problems are often NP-hard and out of reach for exact algorithms.
GFlowNets have emerged as a powerful machinery to efficiently sample from composite unnormalized densities sequentially.
In this paper, we design Markov decision processes (MDPs) for different problems and propose to train conditional GFlowNets to sample from the solution space.
arXiv Detail & Related papers (2023-05-26T15:13:09Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Learning to Control Local Search for Combinatorial Optimization [4.243592852049962]
A zoo of generic as well as problem-specific variants of local search is commonly used to compute approximate solutions.
In this paper we identify three independent algorithmic aspects of such local search algorithms and formalize their sequential selection over an optimization process.
We show that NeuroLS is able to outperform both, well-known general purpose local search controllers from Operations Research as well as latest machine learning-based approaches.
arXiv Detail & Related papers (2022-06-27T10:48:56Z) - Reinforcement Learning for Branch-and-Bound Optimisation using
Retrospective Trajectories [72.15369769265398]
Machine learning has emerged as a promising paradigm for branching.
We propose retro branching; a simple yet effective approach to RL for branching.
We outperform the current state-of-the-art RL branching algorithm by 3-5x and come within 20% of the best IL method's performance on MILPs with 500 constraints and 1000 variables.
arXiv Detail & Related papers (2022-05-28T06:08:07Z) - The Machine Learning for Combinatorial Optimization Competition (ML4CO):
Results and Insights [59.93939636422896]
The ML4CO aims at improving state-of-the-art optimization solvers by replacing key components.
The competition featured three challenging tasks: finding the best feasible solution, producing the tightest optimality certificate, and giving an appropriate routing configuration.
arXiv Detail & Related papers (2022-03-04T17:06:00Z) - Minimizing Entropy to Discover Good Solutions to Recurrent Mixed Integer
Programs [0.0]
Current solvers for mixed-integer programming (MIP) problems are designed to perform well on a wide range of problems.
Recent works have shown that machine learning (ML) can be integrated with an MIP solver to inject domain knowledge and efficiently close the optimality gap.
This paper proposes an online solver that uses the notion of entropy to efficiently build a model with minimal training data and tuning.
arXiv Detail & Related papers (2022-02-07T18:52:56Z) - Meta-aprendizado para otimizacao de parametros de redes neurais [0.9558392439655014]
We investigated the use of meta-learning to the optimization of ANNs.
We performed a case study using meta-learning to choose the number of hidden nodes for networks.
arXiv Detail & Related papers (2021-07-10T15:38:01Z) - POMO: Policy Optimization with Multiple Optima for Reinforcement
Learning [8.819672165548477]
We introduce Policy Optimization with Multiple Optima (POMO), an end-to-end approach for building such a solver.
POMO is applicable to a wide range of CO problems. It is designed to exploit the symmetries in the representation of a CO solution.
We demonstrate the effectiveness of POMO by solving three popular NP-hard problems, namely, traveling salesman (TSP), capacitated vehicle routing (CVRP), and 0-1 knapsack (KP)
arXiv Detail & Related papers (2020-10-30T00:57:50Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.