SaLinA: Sequential Learning of Agents
- URL: http://arxiv.org/abs/2110.07910v1
- Date: Fri, 15 Oct 2021 07:50:35 GMT
- Title: SaLinA: Sequential Learning of Agents
- Authors: Ludovic Denoyer, Alfredo de la Fuente, Song Duong, Jean-Baptiste Gaya,
Pierre-Alexandre Kamienny, Daniel H. Thompson
- Abstract summary: SaLinA is a library that makes implementing complex sequential learning models easy, including reinforcement learning algorithms.
It is built as an extension of PyTorch: algorithms coded with SALINA can be understood in few minutes by PyTorch users and modified easily.
- Score: 13.822224899460656
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: SaLinA is a simple library that makes implementing complex sequential
learning models easy, including reinforcement learning algorithms. It is built
as an extension of PyTorch: algorithms coded with \SALINA{} can be understood
in few minutes by PyTorch users and modified easily. Moreover, SaLinA naturally
works with multiple CPUs and GPUs at train and test time, thus being a good fit
for the large-scale training use cases. In comparison to existing RL libraries,
SaLinA has a very low adoption cost and capture a large variety of settings
(model-based RL, batch RL, hierarchical RL, multi-agent RL, etc.). But SaLinA
does not only target RL practitioners, it aims at providing sequential learning
capabilities to any deep learning programmer.
Related papers
- CleanQRL: Lightweight Single-file Implementations of Quantum Reinforcement Learning Algorithms [2.536162003546062]
CleanQRL is a library that offers single-script implementations of many Quantum Reinforcement Learning algorithms.<n>Our library provides clear and easy to understand scripts that researchers can quickly adapt to their own needs.
arXiv Detail & Related papers (2025-07-10T09:53:39Z) - SeRL: Self-Play Reinforcement Learning for Large Language Models with Limited Data [65.56911325914582]
We propose Self-play Reinforcement Learning (SeRL) to bootstrap Large Language Models (LLMs) training with limited initial data.<n>The proposed SeRL yields results superior to its counterparts and achieves performance on par with those obtained by high-quality data with verifiable rewards.
arXiv Detail & Related papers (2025-05-25T13:28:04Z) - Syllabus: Portable Curricula for Reinforcement Learning Agents [21.20246467152236]
Syllabus is a library for training RL agents with curriculum learning.
Syllabus provides a universal API for curriculum learning algorithms.
We show first examples of curriculum learning in NetHack and Neural MMO.
arXiv Detail & Related papers (2024-11-18T06:22:30Z) - ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL [80.10358123795946]
We develop a framework for building multi-turn RL algorithms for fine-tuning large language models.
Our framework adopts a hierarchical RL approach and runs two RL algorithms in parallel.
Empirically, we find that ArCHer significantly improves efficiency and performance on agent tasks.
arXiv Detail & Related papers (2024-02-29T18:45:56Z) - How Can LLM Guide RL? A Value-Based Approach [68.55316627400683]
Reinforcement learning (RL) has become the de facto standard practice for sequential decision-making problems by improving future acting policies with feedback.
Recent developments in large language models (LLMs) have showcased impressive capabilities in language understanding and generation, yet they fall short in exploration and self-improvement capabilities.
We develop an algorithm named LINVIT that incorporates LLM guidance as a regularization factor in value-based RL, leading to significant reductions in the amount of data needed for learning.
arXiv Detail & Related papers (2024-02-25T20:07:13Z) - StepCoder: Improve Code Generation with Reinforcement Learning from
Compiler Feedback [58.20547418182074]
We introduce StepCoder, a novel framework for code generation, consisting of two main components.
CCCS addresses the exploration challenge by breaking the long sequences code generation task into a Curriculum of Code Completion Subtasks.
FGO only optimize the model by masking the unexecuted code segments to provide Fine-Grained Optimization.
Our method improves the ability to explore the output space and outperforms state-of-the-art approaches in corresponding benchmarks.
arXiv Detail & Related papers (2024-02-02T13:14:31Z) - SRL: Scaling Distributed Reinforcement Learning to Over Ten Thousand Cores [13.948640763797776]
We present a novel abstraction on the dataflows of RL training, which unifies diverse RL training applications into a general framework.
We develop a scalable, efficient, and distributed RL system called ReaLly scalableRL, which allows efficient and massively parallelized training.
SRL is the first in the academic community to perform RL experiments at a large scale with over 15k CPU cores.
arXiv Detail & Related papers (2023-06-29T05:16:25Z) - Is Reinforcement Learning (Not) for Natural Language Processing?:
Benchmarks, Baselines, and Building Blocks for Natural Language Policy
Optimization [73.74371798168642]
We introduce an open-source modular library, RL4LMs, for optimizing language generators with reinforcement learning.
Next, we present the GRUE benchmark, a set of 6 language generation tasks which are supervised not by target strings, but by reward functions.
Finally, we introduce an easy-to-use, performant RL algorithm, NLPO, that learns to effectively reduce the action space in language generation.
arXiv Detail & Related papers (2022-10-03T21:38:29Z) - ShinRL: A Library for Evaluating RL Algorithms from Theoretical and
Practical Perspectives [11.675763847424786]
We present ShinRL, an open-source library for evaluation of reinforcement learning (RL) algorithms.
ShinRL provides an RL environment interface that can compute metrics for delving into the behaviors of RL algorithms.
We show how combining these two features of ShinRL makes it easier to analyze the behavior of deep Q learning.
arXiv Detail & Related papers (2021-12-08T05:34:46Z) - d3rlpy: An Offline Deep Reinforcement Learning Library [0.0]
We introduce d3rlpy, an open-sourced offline deep reinforcement learning (RL) library for Python.
d3rlpy supports a number of offline deep RL algorithms as well as online algorithms via a user-friendly API.
arXiv Detail & Related papers (2021-11-06T03:09:39Z) - Solo-learn: A Library of Self-supervised Methods for Visual
Representation Learning [83.02597612195966]
solo-learn is a library of self-supervised methods for visual representation learning.
Implemented in Python, using Pytorch and Pytorch lightning, the library fits both research and industry needs.
arXiv Detail & Related papers (2021-08-03T22:19:55Z) - TeachMyAgent: a Benchmark for Automatic Curriculum Learning in Deep RL [23.719833581321033]
Training autonomous agents able to generalize to multiple tasks is a key target of Deep Reinforcement Learning (DRL) research.
In parallel to improving DRL algorithms, Automatic Curriculum Learning (ACL) study how teacher algorithms can train DRL agents more efficiently by adapting task selection to their evolving abilities.
While multiple standard benchmarks exist to compare DRL agents, there is currently no such thing for ACL algorithms.
arXiv Detail & Related papers (2021-03-17T17:59:22Z) - MushroomRL: Simplifying Reinforcement Learning Research [60.70556446270147]
MushroomRL is an open-source Python library developed to simplify the process of implementing and running Reinforcement Learning (RL) experiments.
Compared to other available libraries, MushroomRL has been created with the purpose of providing a comprehensive and flexible framework to minimize the effort in implementing and testing novel RL methodologies.
arXiv Detail & Related papers (2020-01-04T17:23:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.