Acme: A Research Framework for Distributed Reinforcement Learning
- URL: http://arxiv.org/abs/2006.00979v2
- Date: Tue, 20 Sep 2022 17:15:51 GMT
- Title: Acme: A Research Framework for Distributed Reinforcement Learning
- Authors: Matthew W. Hoffman, Bobak Shahriari, John Aslanides, Gabriel
Barth-Maron, Nikola Momchev, Danila Sinopalnikov, Piotr Sta\'nczyk, Sabela
Ramos, Anton Raichuk, Damien Vincent, L\'eonard Hussenot, Robert Dadashi,
Gabriel Dulac-Arnold, Manu Orsini, Alexis Jacq, Johan Ferret, Nino Vieillard,
Seyed Kamyar Seyed Ghasemipour, Sertan Girgin, Olivier Pietquin, Feryal
Behbahani, Tamara Norman, Abbas Abdolmaleki, Albin Cassirer, Fan Yang, Kate
Baumli, Sarah Henderson, Abe Friesen, Ruba Haroun, Alex Novikov, Sergio
G\'omez Colmenarejo, Serkan Cabi, Caglar Gulcehre, Tom Le Paine, Srivatsan
Srinivasan, Andrew Cowie, Ziyu Wang, Bilal Piot, Nando de Freitas
- Abstract summary: This paper describes Acme, a framework for constructing novel deep reinforcement learning (RL) algorithms.
It shows how Acme can be used to implement large, distributed RL algorithms that can run at massive scales while still maintaining the inherent readability of that implementation.
- Score: 42.829073211509886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep reinforcement learning (RL) has led to many recent and groundbreaking
advances. However, these advances have often come at the cost of both increased
scale in the underlying architectures being trained as well as increased
complexity of the RL algorithms used to train them. These increases have in
turn made it more difficult for researchers to rapidly prototype new ideas or
reproduce published RL algorithms. To address these concerns this work
describes Acme, a framework for constructing novel RL algorithms that is
specifically designed to enable agents that are built using simple, modular
components that can be used at various scales of execution. While the primary
goal of Acme is to provide a framework for algorithm development, a secondary
goal is to provide simple reference implementations of important or
state-of-the-art algorithms. These implementations serve both as a validation
of our design decisions as well as an important contribution to reproducibility
in RL research. In this work we describe the major design decisions made within
Acme and give further details as to how its components can be used to implement
various algorithms. Our experiments provide baselines for a number of common
and state-of-the-art algorithms as well as showing how these algorithms can be
scaled up for much larger and more complex environments. This highlights one of
the primary advantages of Acme, namely that it can be used to implement large,
distributed RL algorithms that can run at massive scales while still
maintaining the inherent readability of that implementation.
This work presents a second version of the paper which coincides with an
increase in modularity, additional emphasis on offline, imitation and learning
from demonstrations algorithms, as well as various new agents implemented as
part of Acme.
Related papers
- LiGuard: A Streamlined Open-Source Framework for Rapid & Interactive Lidar Research [3.1508266388327324]
LiGuard is an open-source software framework that allows researchers to rapidly develop code for their lidar-based projects.<n>It provides built-in support for data I/O, pre/post processing, and commonly used algorithms.
arXiv Detail & Related papers (2025-09-03T00:10:53Z) - AlphaEvolve: A coding agent for scientific and algorithmic discovery [63.13852052551106]
We present AlphaEvolve, an evolutionary coding agent that substantially enhances capabilities of state-of-the-art LLMs.<n>AlphaEvolve orchestrates an autonomous pipeline of LLMs, whose task is to improve an algorithm by making direct changes to the code.<n>We demonstrate the broad applicability of this approach by applying it to a number of important computational problems.
arXiv Detail & Related papers (2025-06-16T06:37:18Z) - RAGLAB: A Modular and Research-Oriented Unified Framework for Retrieval-Augmented Generation [54.707460684650584]
Large Language Models (LLMs) demonstrate human-level capabilities in dialogue, reasoning, and knowledge retention.
Current research addresses this bottleneck by equipping LLMs with external knowledge, a technique known as Retrieval Augmented Generation (RAG)
RAGLAB is a modular and research-oriented open-source library that reproduces 6 existing algorithms and provides a comprehensive ecosystem for investigating RAG algorithms.
arXiv Detail & Related papers (2024-08-21T07:20:48Z) - Task Agnostic Architecture for Algorithm Induction via Implicit Composition [10.627575117586417]
This position paper aims to explore developing such a unified architecture and proposes a theoretical framework of how it could be constructed.
Recent Generative AI, especially Transformer-based models, demonstrate potential as an architecture capable of constructing algorithms for a wide range of domains.
Our exploration delves into current capabilities and limitations of Transformer-based and other methods in efficient and correct algorithm composition.
arXiv Detail & Related papers (2024-04-03T04:31:09Z) - A Generalist Neural Algorithmic Learner [18.425083543441776]
We build a single graph neural network processor capable of learning to execute a wide range of algorithms.
We show that it is possible to effectively learn algorithms in a multi-task manner, so long as we can learn to execute them well in a single-task regime.
arXiv Detail & Related papers (2022-09-22T16:41:33Z) - A Pragmatic Look at Deep Imitation Learning [0.3626013617212666]
We re-implement 6 different adversarial imitation learning algorithms.
We evaluate them on a widely-used expert trajectory dataset.
GAIL consistently performs well across a range of sample sizes.
arXiv Detail & Related papers (2021-08-04T06:33:10Z) - Identifying Co-Adaptation of Algorithmic and Implementational
Innovations in Deep Reinforcement Learning: A Taxonomy and Case Study of
Inference-based Algorithms [15.338931971492288]
We focus on a series of inference-based actor-critic algorithms to decouple their algorithmic innovations and implementation decisions.
We identify substantial performance drops whenever implementation details are mismatched for algorithmic choices.
Results show which implementation details are co-adapted and co-evolved with algorithms.
arXiv Detail & Related papers (2021-03-31T17:55:20Z) - SUNRISE: A Simple Unified Framework for Ensemble Learning in Deep
Reinforcement Learning [102.78958681141577]
We present SUNRISE, a simple unified ensemble method, which is compatible with various off-policy deep reinforcement learning algorithms.
SUNRISE integrates two key ingredients: (a) ensemble-based weighted Bellman backups, which re-weight target Q-values based on uncertainty estimates from a Q-ensemble, and (b) an inference method that selects actions using the highest upper-confidence bounds for efficient exploration.
arXiv Detail & Related papers (2020-07-09T17:08:44Z) - A Brief Look at Generalization in Visual Meta-Reinforcement Learning [56.50123642237106]
We evaluate the generalization performance of meta-reinforcement learning algorithms.
We find that these algorithms can display strong overfitting when they are evaluated on challenging tasks.
arXiv Detail & Related papers (2020-06-12T15:17:17Z) - Reinforcement Learning with Augmented Data [97.42819506719191]
We present Reinforcement Learning with Augmented Data (RAD), a simple plug-and-play module that can enhance most RL algorithms.
We show that augmentations such as random translate, crop, color jitter, patch cutout, random convolutions, and amplitude scale can enable simple RL algorithms to outperform complex state-of-the-art methods.
arXiv Detail & Related papers (2020-04-30T17:35:32Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.