Sampling-based Pareto Optimization for Chance-constrained Monotone Submodular Problems
- URL: http://arxiv.org/abs/2404.11907v1
- Date: Thu, 18 Apr 2024 05:15:20 GMT
- Title: Sampling-based Pareto Optimization for Chance-constrained Monotone Submodular Problems
- Authors: Xiankun Yan, Aneta Neumann, Frank Neumann,
- Abstract summary: In this paper, a sampling-based method is proposed to directly evaluate the chance constraint.
To address the problems with more challenging settings, an enhanced GSEMO algorithm is introduced.
The ASW-GSEMO with the sampling-based evaluation approach outperforms other algorithms.
- Score: 10.506038775815094
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently surrogate functions based on the tail inequalities were developed to evaluate the chance constraints in the context of evolutionary computation and several Pareto optimization algorithms using these surrogates were successfully applied in optimizing chance-constrained monotone submodular problems. However, the difference in performance between algorithms using the surrogates and those employing the direct sampling-based evaluation remains unclear. Within the paper, a sampling-based method is proposed to directly evaluate the chance constraint. Furthermore, to address the problems with more challenging settings, an enhanced GSEMO algorithm integrated with an adaptive sliding window, called ASW-GSEMO, is introduced. In the experiments, the ASW-GSEMO employing the sampling-based approach is tested on the chance-constrained version of the maximum coverage problem with different settings. Its results are compared with those from other algorithms using different surrogate functions. The experimental findings indicate that the ASW-GSEMO with the sampling-based evaluation approach outperforms other algorithms, highlighting that the performances of algorithms using different evaluation methods are comparable. Additionally, the behaviors of ASW-GSEMO are visualized to explain the distinctions between it and the algorithms utilizing the surrogate functions.
Related papers
- Sliding Window Bi-Objective Evolutionary Algorithms for Optimizing Chance-Constrained Monotone Submodular Functions [10.506038775815094]
In this paper, we apply the sliding-selection approach introduced in [21] to the optimization of chance-constrained monotone submodular functions.
We show that the SW-GSEMO algorithm successfully limits the population size as a key factor that impacts the runtime.
arXiv Detail & Related papers (2024-07-13T00:28:29Z) - Modified LAB Algorithm with Clustering-based Search Space Reduction
Method for solving Engineering Design Problems [0.7789406630452325]
A modified LAB algorithm is introduced in this paper.
The proposed algorithm incorporates the roulette wheel approach and a reduction factor introducing inter-group competition.
The algorithm exhibited improved and superior robustness as well as search space exploration capabilities.
arXiv Detail & Related papers (2023-10-04T12:35:13Z) - Best-Effort Adaptation [62.00856290846247]
We present a new theoretical analysis of sample reweighting methods, including bounds holding uniformly over the weights.
We show how these bounds can guide the design of learning algorithms that we discuss in detail.
We report the results of a series of experiments demonstrating the effectiveness of our best-effort adaptation and domain adaptation algorithms.
arXiv Detail & Related papers (2023-05-10T00:09:07Z) - Exploring the effectiveness of surrogate-assisted evolutionary
algorithms on the batch processing problem [0.0]
This paper introduces a simulation of a well-known batch processing problem in the literature.
Evolutionary algorithms such as Genetic Algorithm (GA), Differential Evolution (DE) are used to find the optimal schedule for the simulation.
We then compare the quality of solutions obtained by the surrogate-assisted versions of the algorithms against the baseline algorithms.
arXiv Detail & Related papers (2022-10-31T09:00:39Z) - An Empirical Evaluation of Zeroth-Order Optimization Methods on
AI-driven Molecule Optimization [78.36413169647408]
We study the effectiveness of various ZO optimization methods for optimizing molecular objectives.
We show the advantages of ZO sign-based gradient descent (ZO-signGD)
We demonstrate the potential effectiveness of ZO optimization methods on widely used benchmark tasks from the Guacamol suite.
arXiv Detail & Related papers (2022-10-27T01:58:10Z) - Amortized Implicit Differentiation for Stochastic Bilevel Optimization [53.12363770169761]
We study a class of algorithms for solving bilevel optimization problems in both deterministic and deterministic settings.
We exploit a warm-start strategy to amortize the estimation of the exact gradient.
By using this framework, our analysis shows these algorithms to match the computational complexity of methods that have access to an unbiased estimate of the gradient.
arXiv Detail & Related papers (2021-11-29T15:10:09Z) - A novel multiobjective evolutionary algorithm based on decomposition and
multi-reference points strategy [14.102326122777475]
Multiobjective evolutionary algorithm based on decomposition (MOEA/D) has been regarded as a significantly promising approach for solving multiobjective optimization problems (MOPs)
We propose an improved MOEA/D algorithm by virtue of the well-known Pascoletti-Serafini scalarization method and a new strategy of multi-reference points.
arXiv Detail & Related papers (2021-10-27T02:07:08Z) - Harnessing Heterogeneity: Learning from Decomposed Feedback in Bayesian
Modeling [68.69431580852535]
We introduce a novel GP regression to incorporate the subgroup feedback.
Our modified regression has provably lower variance -- and thus a more accurate posterior -- compared to previous approaches.
We execute our algorithm on two disparate social problems.
arXiv Detail & Related papers (2021-07-07T03:57:22Z) - Learning Sampling Policy for Faster Derivative Free Optimization [100.27518340593284]
We propose a new reinforcement learning based ZO algorithm (ZO-RL) with learning the sampling policy for generating the perturbations in ZO optimization instead of using random sampling.
Our results show that our ZO-RL algorithm can effectively reduce the variances of ZO gradient by learning a sampling policy, and converge faster than existing ZO algorithms in different scenarios.
arXiv Detail & Related papers (2021-04-09T14:50:59Z) - Optimizing Monotone Chance-Constrained Submodular Functions Using Evolutionary Multi-Objective Algorithms [11.807734722701786]
Here the constraint involves components and the constraint can only be violated with a small probability of alpha.
We show that the algorithm GSEMO obtains the same worst case performance guarantees for monotone submodular functions.
Our experimental results show that the use of evolutionary multi-objective algorithms leads to significant performance improvements.
arXiv Detail & Related papers (2020-06-20T00:17:44Z) - Reparameterized Variational Divergence Minimization for Stable Imitation [57.06909373038396]
We study the extent to which variations in the choice of probabilistic divergence may yield more performant ILO algorithms.
We contribute a re parameterization trick for adversarial imitation learning to alleviate the challenges of the promising $f$-divergence minimization framework.
Empirically, we demonstrate that our design choices allow for ILO algorithms that outperform baseline approaches and more closely match expert performance in low-dimensional continuous-control tasks.
arXiv Detail & Related papers (2020-06-18T19:04:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.