Particle Swarm Optimization: Development of a General-Purpose Optimizer
- URL: http://arxiv.org/abs/2101.09835v1
- Date: Mon, 25 Jan 2021 00:35:18 GMT
- Title: Particle Swarm Optimization: Development of a General-Purpose Optimizer
- Authors: Mauro S. Innocente, Johann Sienz
- Abstract summary: The particle swarm optimization (PSO) method is sometimes viewed as another evolutionary algorithm because of their many similarities.
This paper deals with three important aspects of the method: the influence of the parameters' tuning on the behaviour of the system; the design of stopping criteria so that the reliability of the solution found can be somehow estimated and computational cost can be saved.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Traditional methods present a very restrictive range of applications, mainly
limited by the features of the function to be optimized and of the constraint
functions. In contrast, evolutionary algorithms present almost no restriction
to the features of these functions, although the most appropriate
constraint-handling technique is still an open question. The particle swarm
optimization (PSO) method is sometimes viewed as another evolutionary algorithm
because of their many similarities, despite not being inspired by the same
metaphor. Namely, they evolve a population of individuals taking into
consideration previous experiences and using stochastic operators to introduce
new responses. The advantages of evolutionary algorithms with respect to
traditional methods have been greatly discussed in the literature for decades.
While all such advantages are valid when comparing the PSO paradigm to
traditional methods, its main advantages with respect to evolutionary
algorithms consist of its noticeably lower computational cost and easier
implementation. In fact, the plain version can be programmed in a few lines of
code, involving no operator design and few parameters to be tuned. This paper
deals with three important aspects of the method: the influence of the
parameters' tuning on the behaviour of the system; the design of stopping
criteria so that the reliability of the solution found can be somehow estimated
and computational cost can be saved; and the development of appropriate
techniques to handle constraints, given that the original method is designed
for unconstrained optimization problems.
Related papers
- Learning Joint Models of Prediction and Optimization [56.04498536842065]
Predict-Then-Then framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by joint predictive models.
arXiv Detail & Related papers (2024-09-07T19:52:14Z) - Analyzing and Enhancing the Backward-Pass Convergence of Unrolled
Optimization [50.38518771642365]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
A central challenge in this setting is backpropagation through the solution of an optimization problem, which often lacks a closed form.
This paper provides theoretical insights into the backward pass of unrolled optimization, showing that it is equivalent to the solution of a linear system by a particular iterative method.
A system called Folded Optimization is proposed to construct more efficient backpropagation rules from unrolled solver implementations.
arXiv Detail & Related papers (2023-12-28T23:15:18Z) - Socio-cognitive Optimization of Time-delay Control Problems using
Evolutionary Metaheuristics [89.24951036534168]
Metaheuristics are universal optimization algorithms which should be used for solving difficult problems, unsolvable by classic approaches.
In this paper we aim at constructing novel socio-cognitive metaheuristic based on castes, and apply several versions of this algorithm to optimization of time-delay system model.
arXiv Detail & Related papers (2022-10-23T22:21:10Z) - Generalizing Bayesian Optimization with Decision-theoretic Entropies [102.82152945324381]
We consider a generalization of Shannon entropy from work in statistical decision theory.
We first show that special cases of this entropy lead to popular acquisition functions used in BO procedures.
We then show how alternative choices for the loss yield a flexible family of acquisition functions.
arXiv Detail & Related papers (2022-10-04T04:43:58Z) - Distributed Evolution Strategies for Black-box Stochastic Optimization [42.90600124972943]
This work concerns the evolutionary approaches to distributed black-box optimization.
Each worker can individually solve an approximation of the problem with algorithms.
We propose two alternative simulation schemes which significantly improve robustness of problems.
arXiv Detail & Related papers (2022-04-09T11:18:41Z) - On Constraints in First-Order Optimization: A View from Non-Smooth
Dynamical Systems [99.59934203759754]
We introduce a class of first-order methods for smooth constrained optimization.
Two distinctive features of our approach are that projections or optimizations over the entire feasible set are avoided.
The resulting algorithmic procedure is simple to implement even when constraints are nonlinear.
arXiv Detail & Related papers (2021-07-17T11:45:13Z) - Population-Based Methods: PARTICLE SWARM OPTIMIZATION -- Development of
a General-Purpose Optimizer and Applications [0.0]
This thesis is concerned with continuous, static, and single-objective optimization problems subject to inequality constraints.
The particle swarm optimization paradigm was inspired by previous simulations of the cooperative behaviour observed in social beings.
arXiv Detail & Related papers (2021-01-25T09:36:25Z) - Particle Swarm Optimization: Fundamental Study and its Application to
Optimization and to Jetty Scheduling Problems [0.0]
The advantages of evolutionary algorithms with respect to traditional methods have been greatly discussed in the literature.
While particle swarms share such advantages, they outperform evolutionary algorithms in that they require lower computational cost and easier implementation.
This paper does not intend to study their tuning, general-purpose settings are taken from previous studies, and virtually the same algorithm is used to optimize a variety of notably different problems.
arXiv Detail & Related papers (2021-01-25T02:06:30Z) - Constraint-Handling Techniques for Particle Swarm Optimization
Algorithms [0.0]
Population-based methods can cope with a variety of different problems, including problems of remarkably higher complexity than those traditional methods can handle.
The aim here is to develop and compare different CHTs suitable for PSOs, which are incorporated to an algorithm with general-purpose settings.
arXiv Detail & Related papers (2021-01-25T01:49:10Z) - A Study of the Fundamental Parameters of Particle Swarm Optimizers [0.0]
Population-based algorithms can handle different optimization problems with few or no adaptations.
Their main drawbacks consist of their comparatively higher computational cost and difficulty in handling equality constraints.
This paper deals with the effect of the settings of the parameters of the particles' velocity update equation on the behaviour of the system.
arXiv Detail & Related papers (2021-01-25T01:18:34Z) - Divide and Learn: A Divide and Conquer Approach for Predict+Optimize [50.03608569227359]
The predict+optimize problem combines machine learning ofproblem coefficients with a optimization prob-lem that uses the predicted coefficients.
We show how to directlyexpress the loss of the optimization problem in terms of thepredicted coefficients as a piece-wise linear function.
We propose a novel divide and algorithm to tackle optimization problems without this restriction and predict itscoefficients using the optimization loss.
arXiv Detail & Related papers (2020-12-04T00:26:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.