Recent Trends in the Use of Statistical Tests for Comparing Swarm and
Evolutionary Computing Algorithms: Practical Guidelines and a Critical Review
- URL: http://arxiv.org/abs/2002.09227v1
- Date: Fri, 21 Feb 2020 11:06:47 GMT
- Title: Recent Trends in the Use of Statistical Tests for Comparing Swarm and
Evolutionary Computing Algorithms: Practical Guidelines and a Critical Review
- Authors: J. Carrasco, S. Garc\'ia, M.M. Rueda, S. Das and F. Herrera
- Abstract summary: We conduct a survey on the current trends of the proposals of statistical analyses for the comparison of algorithms of computational intelligence.
We describe the main advantages and drawbacks of the use of each kind of test and put forward some recommendations concerning their use.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A key aspect of the design of evolutionary and swarm intelligence algorithms
is studying their performance. Statistical comparisons are also a crucial part
which allows for reliable conclusions to be drawn. In the present paper we
gather and examine the approaches taken from different perspectives to
summarise the assumptions made by these statistical tests, the conclusions
reached and the steps followed to perform them correctly. In this paper, we
conduct a survey on the current trends of the proposals of statistical analyses
for the comparison of algorithms of computational intelligence and include a
description of the statistical background of these tests. We illustrate the use
of the most common tests in the context of the Competition on single-objective
real parameter optimisation of the IEEE Congress on Evolutionary Computation
(CEC) 2017 and describe the main advantages and drawbacks of the use of each
kind of test and put forward some recommendations concerning their use.
Related papers
- Generalization Bounds of Surrogate Policies for Combinatorial Optimization Problems [61.580419063416734]
A recent stream of structured learning approaches has improved the practical state of the art for a range of optimization problems.
The key idea is to exploit the statistical distribution over instances instead of dealing with instances separately.
In this article, we investigate methods that smooth the risk by perturbing the policy, which eases optimization and improves the generalization error.
arXiv Detail & Related papers (2024-07-24T12:00:30Z) - Globally-Optimal Greedy Experiment Selection for Active Sequential
Estimation [1.1530723302736279]
We study the problem of active sequential estimation, which involves adaptively selecting experiments for sequentially collected data.
The goal is to design experiment selection rules for more accurate model estimation.
We propose a class of greedy experiment selection methods and provide statistical analysis for the maximum likelihood.
arXiv Detail & Related papers (2024-02-13T17:09:29Z) - Measuring Software Testability via Automatically Generated Test Cases [8.17364116624769]
We propose a new approach to pursuing testability measurements based on software metrics.
Our approach exploits automatic test generation and mutation analysis to quantify the evidence about the relative hardness of developing effective test cases.
arXiv Detail & Related papers (2023-07-30T09:48:51Z) - Improved Policy Evaluation for Randomized Trials of Algorithmic Resource
Allocation [54.72195809248172]
We present a new estimator leveraging our proposed novel concept, that involves retrospective reshuffling of participants across experimental arms at the end of an RCT.
We prove theoretically that such an estimator is more accurate than common estimators based on sample means.
arXiv Detail & Related papers (2023-02-06T05:17:22Z) - SELECTOR: Selecting a Representative Benchmark Suite for Reproducible
Statistical Comparison [0.7046417074932257]
We evaluate three approaches for selecting diverse problem instances which should be involved in the comparison of optimization algorithms.
The first approach employs clustering to identify similar groups of problem instances and subsequent sampling from each cluster to construct new benchmarks.
The other two approaches use graph algorithms for identifying dominating and maximal independent sets of nodes.
arXiv Detail & Related papers (2022-04-25T09:38:43Z) - Differential privacy and robust statistics in high dimensions [49.50869296871643]
High-dimensional Propose-Test-Release (HPTR) builds upon three crucial components: the exponential mechanism, robust statistics, and the Propose-Test-Release mechanism.
We show that HPTR nearly achieves the optimal sample complexity under several scenarios studied in the literature.
arXiv Detail & Related papers (2021-11-12T06:36:40Z) - A Field Guide to Federated Optimization [161.3779046812383]
Federated learning and analytics are a distributed approach for collaboratively learning models (or statistics) from decentralized data.
This paper provides recommendations and guidelines on formulating, designing, evaluating and analyzing federated optimization algorithms.
arXiv Detail & Related papers (2021-07-14T18:09:08Z) - More Powerful Conditional Selective Inference for Generalized Lasso by
Parametric Programming [20.309302270008146]
Conditional selective inference (SI) has been studied intensively as a new statistical inference framework for data-driven hypotheses.
We propose a more powerful and general conditional SI method for a class of problems that can be converted into quadratic parametric programming.
arXiv Detail & Related papers (2021-05-11T10:12:00Z) - SAMBA: Safe Model-Based & Active Reinforcement Learning [59.01424351231993]
SAMBA is a framework for safe reinforcement learning that combines aspects from probabilistic modelling, information theory, and statistics.
We evaluate our algorithm on a variety of safe dynamical system benchmarks involving both low and high-dimensional state representations.
We provide intuition as to the effectiveness of the framework by a detailed analysis of our active metrics and safety constraints.
arXiv Detail & Related papers (2020-06-12T10:40:46Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.