A hybrid estimation of distribution algorithm for joint stratification
and sample allocation
- URL: http://arxiv.org/abs/2201.04068v1
- Date: Sun, 9 Jan 2022 21:27:16 GMT
- Title: A hybrid estimation of distribution algorithm for joint stratification
and sample allocation
- Authors: Mervyn O'Luing, Steven Prestwich and S. Armagan Tarim
- Abstract summary: We propose a hybrid estimation of distribution algorithm (HEDA) to solve the joint stratification and sample allocation problem.
EDAs are black-box optimization algorithms which can be used to estimate, build and sample probability models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this study we propose a hybrid estimation of distribution algorithm (HEDA)
to solve the joint stratification and sample allocation problem. This is a
complex problem in which each the quality of each stratification from the set
of all possible stratifications is measured its optimal sample allocation. EDAs
are stochastic black-box optimization algorithms which can be used to estimate,
build and sample probability models in the search for an optimal
stratification. In this paper we enhance the exploitation properties of the EDA
by adding a simulated annealing algorithm to make it a hybrid EDA. Results of
empirical comparisons for atomic and continuous strata show that the HEDA
attains the bests results found so far when compared to benchmark tests on the
same data using a grouping genetic algorithm, simulated annealing algorithm or
hill-climbing algorithm. However, the execution times and total execution are,
in general, higher for the HEDA.
Related papers
- Faster WIND: Accelerating Iterative Best-of-$N$ Distillation for LLM Alignment [81.84950252537618]
This paper reveals a unified game-theoretic connection between iterative BOND and self-play alignment.
We establish a novel framework, WIN rate Dominance (WIND), with a series of efficient algorithms for regularized win rate dominance optimization.
arXiv Detail & Related papers (2024-10-28T04:47:39Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Stability and Generalization of the Decentralized Stochastic Gradient
Descent Ascent Algorithm [80.94861441583275]
We investigate the complexity of the generalization bound of the decentralized gradient descent (D-SGDA) algorithm.
Our results analyze the impact of different top factors on the generalization of D-SGDA.
We also balance it with the generalization to obtain the optimal convex-concave setting.
arXiv Detail & Related papers (2023-10-31T11:27:01Z) - Algorithme EM r\'egularis\'e [0.0]
This paper presents a regularized version of the EM algorithm that efficiently uses prior knowledge to cope with a small sample size.
Experiments on real data highlight the good performance of the proposed algorithm for clustering purposes.
arXiv Detail & Related papers (2023-07-04T23:19:25Z) - Computational Complexity of Sub-linear Convergent Algorithms [0.0]
We will explore how starting with a small sample and then geometrically increasing it, and using the solution of the previous sample to compute the new ERM.
This will solve problems with first-order optimization algorithms of sublinear convergence but with lower computational complexity.
arXiv Detail & Related papers (2022-09-29T05:38:06Z) - Selection of the Most Probable Best [2.1095005405219815]
We consider an expected-value ranking and selection (R&S) problem where all k solutions' simulation outputs depend on a common parameter whose uncertainty can be modeled by a distribution.
We define the most probable best (MPB) to be the solution that has the largest probability of being optimal with respect to the distribution.
We devise a series of algorithms that replace the unknown means in the optimality conditions with their estimates and prove the algorithms' sampling ratios achieve the conditions as the simulation budget increases.
arXiv Detail & Related papers (2022-07-15T15:27:27Z) - General Univariate Estimation-of-Distribution Algorithms [9.853329403413701]
Our general model includes EDAs that are more efficient than the existing ones and these may not be difficult to find as we demonstrate for the OneMax and LeadingOnes benchmarks.
Our unified description of the existing algorithms allows a unified analysis of these; we demonstrate this by providing an analysis of genetic drift.
Our general model also includes EDAs that are more efficient than the existing ones and these may not be difficult to find as we demonstrate for the OneMax and LeadingOnes benchmarks.
arXiv Detail & Related papers (2022-06-22T16:32:04Z) - A Robust and Flexible EM Algorithm for Mixtures of Elliptical
Distributions with Missing Data [71.9573352891936]
This paper tackles the problem of missing data imputation for noisy and non-Gaussian data.
A new EM algorithm is investigated for mixtures of elliptical distributions with the property of handling potential missing data.
Experimental results on synthetic data demonstrate that the proposed algorithm is robust to outliers and can be used with non-Gaussian data.
arXiv Detail & Related papers (2022-01-28T10:01:37Z) - Amortized Implicit Differentiation for Stochastic Bilevel Optimization [53.12363770169761]
We study a class of algorithms for solving bilevel optimization problems in both deterministic and deterministic settings.
We exploit a warm-start strategy to amortize the estimation of the exact gradient.
By using this framework, our analysis shows these algorithms to match the computational complexity of methods that have access to an unbiased estimate of the gradient.
arXiv Detail & Related papers (2021-11-29T15:10:09Z) - Combining K-means type algorithms with Hill Climbing for Joint
Stratification and Sample Allocation Designs [0.0]
This is a sample problem in which we search for the optimal stratification from the set of all possible stratifications of basic strata.
evaluating the cost of each solution is expensive.
We compare the above multi-stage combination of algorithms with three recent algorithms and report the solution costs, evaluation times and training times.
arXiv Detail & Related papers (2021-08-18T08:41:58Z) - Plug-And-Play Learned Gaussian-mixture Approximate Message Passing [71.74028918819046]
We propose a plug-and-play compressed sensing (CS) recovery algorithm suitable for any i.i.d. source prior.
Our algorithm builds upon Borgerding's learned AMP (LAMP), yet significantly improves it by adopting a universal denoising function within the algorithm.
Numerical evaluation shows that the L-GM-AMP algorithm achieves state-of-the-art performance without any knowledge of the source prior.
arXiv Detail & Related papers (2020-11-18T16:40:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.