Implementing a GPU-based parallel MAX-MIN Ant System
- URL: http://arxiv.org/abs/2003.11902v1
- Date: Sat, 18 Jan 2020 14:18:34 GMT
- Title: Implementing a GPU-based parallel MAX-MIN Ant System
- Authors: Rafa{\l} Skinderowicz
- Abstract summary: We discuss a range of novel ideas for improving the GPU-based parallel MMAS implementation.
We show that our MMAS implementation is competitive with state-of-the-art GPU-based and multi-core CPU-based parallel ACO implementations.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The MAX-MIN Ant System (MMAS) is one of the best-known Ant Colony
Optimization (ACO) algorithms proven to be efficient at finding satisfactory
solutions to many difficult combinatorial optimization problems. The slow-down
in Moore's law, and the availability of graphics processing units (GPUs)
capable of conducting general-purpose computations at high speed, has sparked
considerable research efforts into the development of GPU-based ACO
implementations. In this paper, we discuss a range of novel ideas for improving
the GPU-based parallel MMAS implementation, allowing it to better utilize the
computing power offered by two subsequent Nvidia GPU architectures.
Specifically, based on the weighted reservoir sampling algorithm we propose a
novel parallel implementation of the node selection procedure, which is at the
heart of the MMAS and other ACO algorithms. We also present a memory-efficient
implementation of another key-component -- the tabu list structure -- which is
used in the ACO's solution construction stage. The proposed implementations,
combined with the existing approaches, lead to a total of six MMAS variants,
which are evaluated on a set of Traveling Salesman Problem (TSP) instances
ranging from 198 to 3,795 cities. The results show that our MMAS implementation
is competitive with state-of-the-art GPU-based and multi-core CPU-based
parallel ACO implementations: in fact, the times obtained for the Nvidia V100
Volta GPU were up to 7.18x and 21.79x smaller, respectively. The fastest of the
proposed MMAS variants is able to generate over 1 million candidate solutions
per second when solving a 1,002-city instance. Moreover, we show that, combined
with the 2-opt local search heuristic, the proposed parallel MMAS finds
high-quality solutions for the TSP instances with up to 18,512 nodes.
Related papers
- Efficient algorithms to solve atom reconfiguration problems. III. The bird and batching algorithms and other parallel implementations on GPUs [38.2058847337805]
We present efficient implementations of atom reconfiguration algorithms for both CPU and GPU.
Our approach derives improved algorithms that achieve reduced time complexity and faster operational running times.
These algorithms can be used to prepare defect-free configurations of neutral atoms in arrays of optical traps.
arXiv Detail & Related papers (2025-04-08T16:22:42Z) - A GPU Implementation of Multi-Guiding Spark Fireworks Algorithm for Efficient Black-Box Neural Network Optimization [2.9608128305931825]
This paper presents a GPU-accelerated version of the Multi-Guiding Spark Fireworks Algorithm (MGFWA)
We demonstrate its superior performance in terms of both speed and solution quality.
The proposed implementation offers a promising approach to accelerate swarm intelligence algorithms.
arXiv Detail & Related papers (2025-01-07T17:09:07Z) - EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference [49.94169109038806]
This paper introduces EPS-MoE, a novel expert pipeline scheduler for MoE.
Our results demonstrate an average 21% improvement in prefill throughput over existing parallel inference methods.
arXiv Detail & Related papers (2024-10-16T05:17:49Z) - GPU Based Differential Evolution: New Insights and Comparative Study [7.5961910202572644]
This work reviews the main architectural choices made in the literature for GPU based Differential Evolution algorithms.
It introduces a new GPU based numerical optimisation benchmark to evaluate and compare GPU based DE algorithms.
arXiv Detail & Related papers (2024-05-26T12:40:39Z) - An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - New Core-Guided and Hitting Set Algorithms for Multi-Objective
Combinatorial Optimization [0.0]
We present two novel unsatisfiability-based algorithms for Multi-Objective Combinatorial Optimization.
The first is a core-guided MOCO solver, the second is a hitting set-based MOCO solver.
Experimental results show that our new unsatisfiability-based algorithms can outperform state-of-the-art SAT-based algorithms for MOCO.
arXiv Detail & Related papers (2022-04-22T09:46:44Z) - Distributed Out-of-Memory NMF on CPU/GPU Architectures [1.0051474951635875]
We propose an efficient out-of-memory implementation of the Non-negative Matrix Factorization (NMF) algorithm for HPC systems.
Benchmark results show significant improvement of 32X to 76x speedup with the new implementation using GPU over the CPU-based NMFk.
arXiv Detail & Related papers (2022-02-19T03:49:21Z) - A Simple Evolutionary Algorithm for Multi-modal Multi-objective
Optimization [0.0]
We introduce a steady-state evolutionary algorithm for solving multi-modal, multi-objective optimization problems (MMOPs)
We report its performance on 21 MMOPs from various test suites that are widely used for benchmarking using a low computational budget of 1000 function evaluations.
arXiv Detail & Related papers (2022-01-18T03:31:11Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - Kernel methods through the roof: handling billions of points efficiently [94.31450736250918]
Kernel methods provide an elegant and principled approach to nonparametric learning, but so far could hardly be used in large scale problems.
Recent advances have shown the benefits of a number of algorithmic ideas, for example combining optimization, numerical linear algebra and random projections.
Here, we push these efforts further to develop and test a solver that takes full advantage of GPU hardware.
arXiv Detail & Related papers (2020-06-18T08:16:25Z) - Iterative Algorithm Induced Deep-Unfolding Neural Networks: Precoding
Design for Multiuser MIMO Systems [59.804810122136345]
We propose a framework for deep-unfolding, where a general form of iterative algorithm induced deep-unfolding neural network (IAIDNN) is developed.
An efficient IAIDNN based on the structure of the classic weighted minimum mean-square error (WMMSE) iterative algorithm is developed.
We show that the proposed IAIDNN efficiently achieves the performance of the iterative WMMSE algorithm with reduced computational complexity.
arXiv Detail & Related papers (2020-06-15T02:57:57Z) - MPLP++: Fast, Parallel Dual Block-Coordinate Ascent for Dense Graphical
Models [96.1052289276254]
This work introduces a new MAP-solver, based on the popular Dual Block-Coordinate Ascent principle.
Surprisingly, by making a small change to the low-performing solver, we derive the new solver MPLP++ that significantly outperforms all existing solvers by a large margin.
arXiv Detail & Related papers (2020-04-16T16:20:53Z) - Minimal Filtering Algorithms for Convolutional Neural Networks [82.24592140096622]
We develop fully parallel hardware-oriented algorithms for implementing the basic filtering operation for M=3,5,7,9, and 11.
A fully parallel hardware implementation of the proposed algorithms in each case gives approximately 30 percent savings in the number of embedded multipliers.
arXiv Detail & Related papers (2020-04-12T13:18:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.