Introductory Studies of Swarm Intelligence Techniques
- URL: http://arxiv.org/abs/2209.12823v1
- Date: Mon, 26 Sep 2022 16:29:55 GMT
- Title: Introductory Studies of Swarm Intelligence Techniques
- Authors: Thounaojam Chinglemba, Soujanyo Biswas, Debashish Malakar, Vivek
Meena, Debojyoti Sarkar, and Anupam Biswas
- Abstract summary: Swarm intelligence involves the collective study of individuals and their mutual interactions leading to intelligent behavior of the swarm.
The chapter presents various population-based SI algorithms, their fundamental structures along with their mathematical models.
- Score: 1.2930503923129208
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rapid upliftment of technology, there has emerged a dire need to
fine-tune or optimize certain processes, software, models or structures, with
utmost accuracy and efficiency. Optimization algorithms are preferred over
other methods of optimization through experimentation or simulation, for their
generic problem-solving abilities and promising efficacy with the least human
intervention. In recent times, the inducement of natural phenomena into
algorithm design has immensely triggered the efficiency of optimization process
for even complex multi-dimensional, non-continuous, non-differentiable and
noisy problem search spaces. This chapter deals with the Swarm intelligence
(SI) based algorithms or Swarm Optimization Algorithms, which are a subset of
the greater Nature Inspired Optimization Algorithms (NIOAs). Swarm intelligence
involves the collective study of individuals and their mutual interactions
leading to intelligent behavior of the swarm. The chapter presents various
population-based SI algorithms, their fundamental structures along with their
mathematical models.
Related papers
- Nature-Inspired Algorithms in Optimization: Introduction, Hybridization
and Insights [1.6589012298747952]
Benchmarking is important in evaluating the performance of optimization algorithms.
This chapter focuses on the overview of optimization, nature-inspired algorithms and the role of hybridization.
arXiv Detail & Related papers (2023-08-30T11:33:22Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Enhancing Machine Learning Model Performance with Hyper Parameter
Optimization: A Comparative Study [0.0]
One of the most critical issues in machine learning is the selection of appropriate hyper parameters for training models.
Hyper parameter optimization (HPO) is a popular topic that artificial intelligence studies have focused on recently.
In this study, classical methods, such as grid, random search and Bayesian optimization, and population-based algorithms, such as genetic algorithms and particle swarm optimization, are discussed.
arXiv Detail & Related papers (2023-02-14T10:12:10Z) - Socio-cognitive Optimization of Time-delay Control Problems using
Evolutionary Metaheuristics [89.24951036534168]
Metaheuristics are universal optimization algorithms which should be used for solving difficult problems, unsolvable by classic approaches.
In this paper we aim at constructing novel socio-cognitive metaheuristic based on castes, and apply several versions of this algorithm to optimization of time-delay system model.
arXiv Detail & Related papers (2022-10-23T22:21:10Z) - Neural Combinatorial Optimization: a New Player in the Field [69.23334811890919]
This paper presents a critical analysis on the incorporation of algorithms based on neural networks into the classical optimization framework.
A comprehensive study is carried out to analyse the fundamental aspects of such algorithms, including performance, transferability, computational cost and to larger-sized instances.
arXiv Detail & Related papers (2022-05-03T07:54:56Z) - OptABC: an Optimal Hyperparameter Tuning Approach for Machine Learning
Algorithms [1.6114012813668934]
OptABC is proposed to help ABC algorithm in faster convergence toward a near-optimum solution.
OptABC integrates artificial bee colony algorithm, K-Means clustering, greedy algorithm, and opposition-based learning strategy.
Experimental results demonstrate the effectiveness of OptABC compared to existing approaches in the literature.
arXiv Detail & Related papers (2021-12-15T22:33:39Z) - Amortized Implicit Differentiation for Stochastic Bilevel Optimization [53.12363770169761]
We study a class of algorithms for solving bilevel optimization problems in both deterministic and deterministic settings.
We exploit a warm-start strategy to amortize the estimation of the exact gradient.
By using this framework, our analysis shows these algorithms to match the computational complexity of methods that have access to an unbiased estimate of the gradient.
arXiv Detail & Related papers (2021-11-29T15:10:09Z) - Particle Swarm Optimization: Fundamental Study and its Application to
Optimization and to Jetty Scheduling Problems [0.0]
The advantages of evolutionary algorithms with respect to traditional methods have been greatly discussed in the literature.
While particle swarms share such advantages, they outperform evolutionary algorithms in that they require lower computational cost and easier implementation.
This paper does not intend to study their tuning, general-purpose settings are taken from previous studies, and virtually the same algorithm is used to optimize a variety of notably different problems.
arXiv Detail & Related papers (2021-01-25T02:06:30Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Benchmarking Meta-heuristic Optimization [0.0]
Many meta-heuristic algorithms are very efficient when solving nonlinear functions.
A meta-heuristic algorithm is a problem-independent technique that can be applied to a broad range of problems.
arXiv Detail & Related papers (2020-07-27T12:25:31Z) - EOS: a Parallel, Self-Adaptive, Multi-Population Evolutionary Algorithm
for Constrained Global Optimization [68.8204255655161]
EOS is a global optimization algorithm for constrained and unconstrained problems of real-valued variables.
It implements a number of improvements to the well-known Differential Evolution (DE) algorithm.
Results prove that EOSis capable of achieving increased performance compared to state-of-the-art single-population self-adaptive DE algorithms.
arXiv Detail & Related papers (2020-07-09T10:19:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.