Attention-oriented Brain Storm Optimization for Multimodal Optimization
Problems
- URL: http://arxiv.org/abs/2105.13095v1
- Date: Thu, 27 May 2021 12:47:57 GMT
- Title: Attention-oriented Brain Storm Optimization for Multimodal Optimization
Problems
- Authors: Jian Yang and Yuhui Shi
- Abstract summary: This paper presents an Attention-oriented Brain Storm Optimization (ABSO) method that introduces the attention mechanism into a relatively new swarm intelligence algorithm.
Rather than converge to a single global optimum, the proposed method can guide the search procedure to converge to multiple "salient" solutions.
- Score: 24.38312890501329
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Population-based methods are often used to solve multimodal optimization
problems. By combining niching or clustering strategy, the state-of-the-art
approaches generally divide the population into several subpopulations to find
multiple solutions for a problem at hand. However, these methods only guided by
the fitness value during iterations, which are suffering from determining the
number of subpopulations, i.e., the number of niche areas or clusters. To
compensate for this drawback, this paper presents an Attention-oriented Brain
Storm Optimization (ABSO) method that introduces the attention mechanism into a
relatively new swarm intelligence algorithm, i.e., Brain Storm Optimization
(BSO). By converting the objective space from the fitness space into
"attention" space, the individuals are clustered and updated iteratively
according to their salient values. Rather than converge to a single global
optimum, the proposed method can guide the search procedure to converge to
multiple "salient" solutions. The preliminary results show that the proposed
method can locate multiple global and local optimal solutions of several
multimodal benchmark functions. The proposed method needs less prior knowledge
of the problem and can automatically converge to multiple optimums guided by
the attention mechanism, which has excellent potential for further development.
Related papers
- Un-evaluated Solutions May Be Valuable in Expensive Optimization [5.6787965501364335]
We propose a strategic approach that incorporates high-quality, un-evaluated solutions predicted by surrogate models during the selection phase.
This approach aims to improve the distribution of evaluated solutions, thereby generating a superior next generation of solutions.
arXiv Detail & Related papers (2024-12-05T04:06:30Z) - Integrating Chaotic Evolutionary and Local Search Techniques in Decision Space for Enhanced Evolutionary Multi-Objective Optimization [1.8130068086063336]
This paper focuses on both Single-Objective Multi-Modal Optimization (SOMMOP) and Multi-Objective Optimization (MOO)
In SOMMOP, we integrate chaotic evolution with niching techniques, as well as Persistence-Based Clustering combined with Gaussian mutation.
For MOO, we extend these methods into a comprehensive framework that incorporates Uncertainty-Based Selection, Adaptive Tuning, and introduces a radius ( R ) concept in deterministic crowding.
arXiv Detail & Related papers (2024-11-12T15:18:48Z) - Learning Multiple Initial Solutions to Optimization Problems [52.9380464408756]
Sequentially solving similar optimization problems under strict runtime constraints is essential for many applications.
We propose learning to predict emphmultiple diverse initial solutions given parameters that define the problem instance.
We find significant and consistent improvement with our method across all evaluation settings and demonstrate that it efficiently scales with the number of initial solutions required.
arXiv Detail & Related papers (2024-11-04T15:17:19Z) - Training Greedy Policy for Proposal Batch Selection in Expensive Multi-Objective Combinatorial Optimization [52.80408805368928]
We introduce a novel greedy-style subset selection algorithm for batch acquisition.
Our experiments on the red fluorescent proteins show that our proposed method achieves the baseline performance in 1.69x fewer queries.
arXiv Detail & Related papers (2024-06-21T05:57:08Z) - Federated Multi-Level Optimization over Decentralized Networks [55.776919718214224]
We study the problem of distributed multi-level optimization over a network, where agents can only communicate with their immediate neighbors.
We propose a novel gossip-based distributed multi-level optimization algorithm that enables networked agents to solve optimization problems at different levels in a single timescale.
Our algorithm achieves optimal sample complexity, scaling linearly with the network size, and demonstrates state-of-the-art performance on various applications.
arXiv Detail & Related papers (2023-10-10T00:21:10Z) - Optimizing Solution-Samplers for Combinatorial Problems: The Landscape
of Policy-Gradient Methods [52.0617030129699]
We introduce a novel theoretical framework for analyzing the effectiveness of DeepMatching Networks and Reinforcement Learning methods.
Our main contribution holds for a broad class of problems including Max-and Min-Cut, Max-$k$-Bipartite-Bi, Maximum-Weight-Bipartite-Bi, and Traveling Salesman Problem.
As a byproduct of our analysis we introduce a novel regularization process over vanilla descent and provide theoretical and experimental evidence that it helps address vanishing-gradient issues and escape bad stationary points.
arXiv Detail & Related papers (2023-10-08T23:39:38Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Enhanced Opposition Differential Evolution Algorithm for Multimodal
Optimization [0.2538209532048866]
Most of the real-world problems are multimodal in nature that consists of multiple optimum values.
Classical gradient-based methods fail for optimization problems in which the objective functions are either discontinuous or non-differentiable.
We have proposed an algorithm known as Enhanced Opposition Differential Evolution (EODE) algorithm to solve the MMOPs.
arXiv Detail & Related papers (2022-08-23T16:18:27Z) - ARES: An Efficient Algorithm with Recurrent Evaluation and Sampling-Driven Inference for Maximum Independent Set [48.57120672468062]
This paper introduces an efficient algorithm for the Maximum Independent Set (MIS) problem, incorporating two innovative techniques.
The proposed algorithm outperforms state-of-the-art algorithms in terms of solution quality, computational efficiency, and stability.
arXiv Detail & Related papers (2022-08-16T14:39:38Z) - Quant-BnB: A Scalable Branch-and-Bound Method for Optimal Decision Trees
with Continuous Features [5.663538370244174]
We present a new discrete optimization method based on branch-and-bound (BnB) to obtain optimal decision trees.
Our proposed algorithm Quant-BnB shows significant speedups compared to existing approaches for shallow optimal trees on various real datasets.
arXiv Detail & Related papers (2022-06-23T17:19:29Z) - GPSAF: A Generalized Probabilistic Surrogate-Assisted Framework for
Constrained Single- and Multi-objective Optimization [7.8140593450932965]
This paper proposes a generalized probabilistic surrogate-assisted framework (GPSAF)
GPSAF is applicable to a broad category of unconstrained and constrained, single- and multi-objective optimization algorithms.
arXiv Detail & Related papers (2022-04-06T13:22:30Z) - Learning Proximal Operators to Discover Multiple Optima [66.98045013486794]
We present an end-to-end method to learn the proximal operator across non-family problems.
We show that for weakly-ized objectives and under mild conditions, the method converges globally.
arXiv Detail & Related papers (2022-01-28T05:53:28Z) - Population-Based Methods: PARTICLE SWARM OPTIMIZATION -- Development of
a General-Purpose Optimizer and Applications [0.0]
This thesis is concerned with continuous, static, and single-objective optimization problems subject to inequality constraints.
The particle swarm optimization paradigm was inspired by previous simulations of the cooperative behaviour observed in social beings.
arXiv Detail & Related papers (2021-01-25T09:36:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.