Efficient Quality-Diversity Optimization through Diverse Quality Species
- URL: http://arxiv.org/abs/2304.07425v1
- Date: Fri, 14 Apr 2023 23:15:51 GMT
- Title: Efficient Quality-Diversity Optimization through Diverse Quality Species
- Authors: Ryan Wickman, Bibek Poudel, Michael Villarreal, Xiaofei Zhang, Weizi
Li
- Abstract summary: We show that a diverse population of solutions can be found without the limitation of needing an archive or defining the range of behaviors in advance.
We propose Diverse Quality Species (DQS) as an alternative to archive-based Quality-Diversity (QD) algorithms.
- Score: 3.428706362109921
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A prevalent limitation of optimizing over a single objective is that it can
be misguided, becoming trapped in local optimum. This can be rectified by
Quality-Diversity (QD) algorithms, where a population of high-quality and
diverse solutions to a problem is preferred. Most conventional QD approaches,
for example, MAP-Elites, explicitly manage a behavioral archive where solutions
are broken down into predefined niches. In this work, we show that a diverse
population of solutions can be found without the limitation of needing an
archive or defining the range of behaviors in advance. Instead, we break down
solutions into independently evolving species and use unsupervised skill
discovery to learn diverse, high-performing solutions. We show that this can be
done through gradient-based mutations that take on an information theoretic
perspective of jointly maximizing mutual information and performance. We
propose Diverse Quality Species (DQS) as an alternative to archive-based QD
algorithms. We evaluate it over several simulated robotic environments and show
that it can learn a diverse set of solutions from varying species. Furthermore,
our results show that DQS is more sample-efficient and performant when compared
to other QD algorithms. Relevant code and hyper-parameters are available at:
https://github.com/rwickman/NEAT_RL.
Related papers
- Quality-Diversity Algorithms Can Provably Be Helpful for Optimization [24.694984679399315]
Quality-Diversity (QD) algorithms aim to find a set of high-performing, yet diverse solutions.
This paper tries to shed some light on the optimization ability of QD algorithms via rigorous running time analysis.
arXiv Detail & Related papers (2024-01-19T07:40:24Z) - Don't Bet on Luck Alone: Enhancing Behavioral Reproducibility of
Quality-Diversity Solutions in Uncertain Domains [2.639902239625779]
We introduce Archive Reproducibility Improvement Algorithm (ARIA)
ARIA is a plug-and-play approach that improves the quality of solutions present in an archive.
We show that our algorithm enhances the quality and descriptor space coverage of any given archive by at least 50%.
arXiv Detail & Related papers (2023-04-07T14:45:14Z) - Learning to Walk Autonomously via Reset-Free Quality-Diversity [73.08073762433376]
Quality-Diversity algorithms can discover large and complex behavioural repertoires consisting of both diverse and high-performing skills.
Existing QD algorithms need large numbers of evaluations as well as episodic resets, which require manual human supervision and interventions.
This paper proposes Reset-Free Quality-Diversity optimization (RF-QD) as a step towards autonomous learning for robotics in open-ended environments.
arXiv Detail & Related papers (2022-04-07T14:07:51Z) - Approximating Gradients for Differentiable Quality Diversity in
Reinforcement Learning [8.591356221688773]
Differentiable quality diversity (DQD) algorithms greatly accelerate QD optimization when exact gradients are available for the objective and measures.
We develop two variants of the DQD algorithm CMA-MEGA, each with different gradient approximations, and evaluate them on four simulated walking tasks.
One variant achieves comparable performance (QD score) with the state-of-the-art PGA-MAP-Elites in two tasks. The other variant performs comparably in all tasks but is less efficient than PGA-MAP-Elites in two tasks.
arXiv Detail & Related papers (2022-02-08T05:53:55Z) - Result Diversification by Multi-objective Evolutionary Algorithms with
Theoretical Guarantees [94.72461292387146]
We propose to reformulate the result diversification problem as a bi-objective search problem, and solve it by a multi-objective evolutionary algorithm (EA)
We theoretically prove that the GSEMO can achieve the optimal-time approximation ratio, $1/2$.
When the objective function changes dynamically, the GSEMO can maintain this approximation ratio in running time, addressing the open question proposed by Borodin et al.
arXiv Detail & Related papers (2021-10-18T14:00:22Z) - Few-shot Quality-Diversity Optimization [50.337225556491774]
Quality-Diversity (QD) optimization has been shown to be effective tools in dealing with deceptive minima and sparse rewards in Reinforcement Learning.
We show that, given examples from a task distribution, information about the paths taken by optimization in parameter space can be leveraged to build a prior population, which when used to initialize QD methods in unseen environments, allows for few-shot adaptation.
Experiments carried in both sparse and dense reward settings using robotic manipulation and navigation benchmarks show that it considerably reduces the number of generations that are required for QD optimization in these environments.
arXiv Detail & Related papers (2021-09-14T17:12:20Z) - Differentiable Quality Diversity [1.0965065178451106]
We present the differentiable quality diversity (DQD) problem, where both the objective and measure functions are first order differentiable.
We then present MAP-Elites via Gradient Arborescence (MEGA), a DQD algorithm that leverages gradient information to efficiently explore the joint range of the objective and measure functions.
Results in two QD benchmark domains and in searching the latent space of a StyleGAN show that MEGA significantly outperforms state-of-the-art QD algorithms.
arXiv Detail & Related papers (2021-06-07T18:11:53Z) - Ridge Rider: Finding Diverse Solutions by Following Eigenvectors of the
Hessian [48.61341260604871]
Gradient Descent (SGD) is a key component of the success of deep neural networks (DNNs)
In this paper, we present a different approach by following the eigenvectors of the Hessian, which we call "ridges"
We show both theoretically and experimentally that our method, called Ridge Rider (RR), offers a promising direction for a variety of challenging problems.
arXiv Detail & Related papers (2020-11-12T17:15:09Z) - A Discriminative Technique for Multiple-Source Adaptation [55.5865665284915]
We present a new discriminative technique for the multiple-source adaptation, MSA, problem.
Our solution only requires conditional probabilities that can easily be accurately estimated from unlabeled data from the source domains.
Our experiments with real-world applications further demonstrate that our new discriminative MSA algorithm outperforms the previous generative solution.
arXiv Detail & Related papers (2020-08-25T14:06:15Z) - BOP-Elites, a Bayesian Optimisation algorithm for Quality-Diversity
search [0.0]
We propose the Bayesian optimisation of Elites (BOP-Elites) algorithm.
By considering user defined regions of the feature space as 'niches' our task is to find the optimal solution in each niche.
The resulting algorithm is very effective in identifying the parts of the search space that belong to a niche in feature space, and finding the optimal solution in each niche.
arXiv Detail & Related papers (2020-05-08T23:49:13Z) - Discovering Representations for Black-box Optimization [73.59962178534361]
We show that black-box optimization encodings can be automatically learned, rather than hand designed.
We show that learned representations make it possible to solve high-dimensional problems with orders of magnitude fewer evaluations than the standard MAP-Elites.
arXiv Detail & Related papers (2020-03-09T20:06:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.