Particle Swarm Optimization: Fundamental Study and its Application to
Optimization and to Jetty Scheduling Problems
- URL: http://arxiv.org/abs/2101.11096v1
- Date: Mon, 25 Jan 2021 02:06:30 GMT
- Title: Particle Swarm Optimization: Fundamental Study and its Application to
Optimization and to Jetty Scheduling Problems
- Authors: Johann Sienz, Mauro S. Innocente
- Abstract summary: The advantages of evolutionary algorithms with respect to traditional methods have been greatly discussed in the literature.
While particle swarms share such advantages, they outperform evolutionary algorithms in that they require lower computational cost and easier implementation.
This paper does not intend to study their tuning, general-purpose settings are taken from previous studies, and virtually the same algorithm is used to optimize a variety of notably different problems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The advantages of evolutionary algorithms with respect to traditional methods
have been greatly discussed in the literature. While particle swarm optimizers
share such advantages, they outperform evolutionary algorithms in that they
require lower computational cost and easier implementation, involving no
operator design and few coefficients to be tuned. However, even marginal
variations in the settings of these coefficients greatly influence the dynamics
of the swarm. Since this paper does not intend to study their tuning,
general-purpose settings are taken from previous studies, and virtually the
same algorithm is used to optimize a variety of notably different problems.
Thus, following a review of the paradigm, the algorithm is tested on a set of
benchmark functions and engineering problems taken from the literature. Later,
complementary lines of code are incorporated to adapt the method to
combinatorial optimization as it occurs in scheduling problems, and a real case
is solved using the same optimizer with the same settings. The aim is to show
the flexibility and robustness of the approach, which can handle a wide variety
of problems.
Related papers
- Learning Joint Models of Prediction and Optimization [56.04498536842065]
Predict-Then-Then framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by joint predictive models.
arXiv Detail & Related papers (2024-09-07T19:52:14Z) - Analyzing and Enhancing the Backward-Pass Convergence of Unrolled
Optimization [50.38518771642365]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
A central challenge in this setting is backpropagation through the solution of an optimization problem, which often lacks a closed form.
This paper provides theoretical insights into the backward pass of unrolled optimization, showing that it is equivalent to the solution of a linear system by a particular iterative method.
A system called Folded Optimization is proposed to construct more efficient backpropagation rules from unrolled solver implementations.
arXiv Detail & Related papers (2023-12-28T23:15:18Z) - Backpropagation of Unrolled Solvers with Folded Optimization [55.04219793298687]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
One typical strategy is algorithm unrolling, which relies on automatic differentiation through the operations of an iterative solver.
This paper provides theoretical insights into the backward pass of unrolled optimization, leading to a system for generating efficiently solvable analytical models of backpropagation.
arXiv Detail & Related papers (2023-01-28T01:50:42Z) - Socio-cognitive Optimization of Time-delay Control Problems using
Evolutionary Metaheuristics [89.24951036534168]
Metaheuristics are universal optimization algorithms which should be used for solving difficult problems, unsolvable by classic approaches.
In this paper we aim at constructing novel socio-cognitive metaheuristic based on castes, and apply several versions of this algorithm to optimization of time-delay system model.
arXiv Detail & Related papers (2022-10-23T22:21:10Z) - Surrogate-based optimization for variational quantum algorithms [0.0]
Variational quantum algorithms are a class of techniques intended to be used on near-term quantum computers.
We introduce the idea of learning surrogate models for variational circuits using few experimental measurements.
We then perform parameter optimization using these models as opposed to the original data.
arXiv Detail & Related papers (2022-04-12T00:15:17Z) - Quantum variational optimization: The role of entanglement and problem
hardness [0.0]
We study the role of entanglement, the structure of the variational quantum circuit, and the structure of the optimization problem.
Our numerical results indicate an advantage in adapting the distribution of entangling gates to the problem's topology.
We find evidence that applying conditional value at risk type cost functions improves the optimization, increasing the probability of overlap with the optimal solutions.
arXiv Detail & Related papers (2021-03-26T14:06:54Z) - A Study of the Fundamental Parameters of Particle Swarm Optimizers [0.0]
Population-based algorithms can handle different optimization problems with few or no adaptations.
Their main drawbacks consist of their comparatively higher computational cost and difficulty in handling equality constraints.
This paper deals with the effect of the settings of the parameters of the particles' velocity update equation on the behaviour of the system.
arXiv Detail & Related papers (2021-01-25T01:18:34Z) - Particle Swarm Optimization: Development of a General-Purpose Optimizer [0.0]
The particle swarm optimization (PSO) method is sometimes viewed as another evolutionary algorithm because of their many similarities.
This paper deals with three important aspects of the method: the influence of the parameters' tuning on the behaviour of the system; the design of stopping criteria so that the reliability of the solution found can be somehow estimated and computational cost can be saved.
arXiv Detail & Related papers (2021-01-25T00:35:18Z) - Divide and Learn: A Divide and Conquer Approach for Predict+Optimize [50.03608569227359]
The predict+optimize problem combines machine learning ofproblem coefficients with a optimization prob-lem that uses the predicted coefficients.
We show how to directlyexpress the loss of the optimization problem in terms of thepredicted coefficients as a piece-wise linear function.
We propose a novel divide and algorithm to tackle optimization problems without this restriction and predict itscoefficients using the optimization loss.
arXiv Detail & Related papers (2020-12-04T00:26:56Z) - Using models to improve optimizers for variational quantum algorithms [1.7475326826331605]
Variational quantum algorithms are a leading candidate for early applications on noisy intermediate-scale quantum computers.
These algorithms depend on a classical optimization outer-loop that minimizes some function of a parameterized quantum circuit.
We introduce two optimization methods and numerically compare their performance with common methods in use today.
arXiv Detail & Related papers (2020-05-22T05:23:23Z) - Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization [71.03797261151605]
Adaptivity is an important yet under-studied property in modern optimization theory.
Our algorithm is proved to achieve the best-available convergence for non-PL objectives simultaneously while outperforming existing algorithms for PL objectives.
arXiv Detail & Related papers (2020-02-13T05:42:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.