An Edge-Aware Graph Autoencoder Trained on Scale-Imbalanced Data for
Traveling Salesman Problems
- URL: http://arxiv.org/abs/2310.06543v2
- Date: Sat, 2 Mar 2024 07:53:46 GMT
- Title: An Edge-Aware Graph Autoencoder Trained on Scale-Imbalanced Data for
Traveling Salesman Problems
- Authors: Shiqing Liu, Xueming Yan, Yaochu Jin
- Abstract summary: This work proposes a data-driven graph representation learning method for solving the Traveling Salesman Problem (TSP)
A residual gated encoder is trained to learn latent edge embeddings, followed by an edge-centered decoder to output link predictions in an end-to-end manner.
The experimental results demonstrate that the proposed edge-aware graph autoencoder model achieves a highly competitive performance.
- Score: 22.792870849003137
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, there has been a notable surge in research on machine
learning techniques for combinatorial optimization. It has been shown that
learning-based methods outperform traditional heuristics and mathematical
solvers on the Traveling Salesman Problem (TSP) in terms of both performance
and computational efficiency. However, most learning-based TSP solvers are
primarily designed for fixed-scale TSP instances, and also require a large
number of training samples to achieve optimal performance. To fill this gap,
this work proposes a data-driven graph representation learning method for
solving TSPs with various numbers of cities. Specifically, we formulate the TSP
as a link prediction task and propose an edge-aware graph autoencoder (EdgeGAE)
model that can solve TSPs by learning from various-scale samples with an
imbalanced distribution. A residual gated encoder is trained to learn latent
edge embeddings, followed by an edge-centered decoder to output link
predictions in an end-to-end manner. Furthermore, we introduce an active
sampling strategy into the training process to improve the model's
generalization capability in large-scale scenarios. To investigate the model's
practical applicability, we generate a scale-imbalanced dataset comprising
50,000 TSP instances ranging from 50 to 500 cities. The experimental results
demonstrate that the proposed edge-aware graph autoencoder model achieves a
highly competitive performance among state-of-the-art graph learning-based
approaches in solving TSPs with various scales, implying its remarkable
potential in dealing with practical optimization challenges.
Related papers
- Novel Representation Learning Technique using Graphs for Performance
Analytics [0.0]
We propose a novel idea of transforming performance data into graphs to leverage the advancement of Graph Neural Network-based (GNN) techniques.
In contrast to other Machine Learning application domains, such as social networks, the graph is not given; instead, we need to build it.
We evaluate the effectiveness of the generated embeddings from GNNs based on how well they make even a simple feed-forward neural network perform for regression tasks.
arXiv Detail & Related papers (2024-01-19T16:34:37Z) - When Parameter-efficient Tuning Meets General-purpose Vision-language
Models [65.19127815275307]
PETAL revolutionizes the training process by requiring only 0.5% of the total parameters, achieved through a unique mode approximation technique.
Our experiments reveal that PETAL not only outperforms current state-of-the-art methods in most scenarios but also surpasses full fine-tuning models in effectiveness.
arXiv Detail & Related papers (2023-12-16T17:13:08Z) - Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Unifying Synergies between Self-supervised Learning and Dynamic
Computation [53.66628188936682]
We present a novel perspective on the interplay between SSL and DC paradigms.
We show that it is feasible to simultaneously learn a dense and gated sub-network from scratch in a SSL setting.
The co-evolution during pre-training of both dense and gated encoder offers a good accuracy-efficiency trade-off.
arXiv Detail & Related papers (2023-01-22T17:12:58Z) - Offline Q-Learning on Diverse Multi-Task Data Both Scales And
Generalizes [100.69714600180895]
offline Q-learning algorithms exhibit strong performance that scales with model capacity.
We train a single policy on 40 games with near-human performance using up-to 80 million parameter networks.
Compared to return-conditioned supervised approaches, offline Q-learning scales similarly with model capacity and has better performance, especially when the dataset is suboptimal.
arXiv Detail & Related papers (2022-11-28T08:56:42Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Learning to Sparsify Travelling Salesman Problem Instances [0.5985204759362747]
We use a pruning machine learning as a pre-processing step followed by an exact Programming approach to sparsify the travelling salesman problem.
Our learning approach requires very little training data and is amenable to mathematical analysis.
arXiv Detail & Related papers (2021-04-19T14:35:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.