NLPGym -- A toolkit for evaluating RL agents on Natural Language
Processing Tasks
- URL: http://arxiv.org/abs/2011.08272v1
- Date: Mon, 16 Nov 2020 20:58:35 GMT
- Title: NLPGym -- A toolkit for evaluating RL agents on Natural Language
Processing Tasks
- Authors: Rajkumar Ramamurthy, Rafet Sifa and Christian Bauckhage
- Abstract summary: We release NLPGym, an open-source Python toolkit that provides interactive textual environments for standard NLP tasks.
We present experimental results for 6 tasks using different RL algorithms which serve as baselines for further research.
- Score: 2.5760935151452067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL) has recently shown impressive performance in
complex game AI and robotics tasks. To a large extent, this is thanks to the
availability of simulated environments such as OpenAI Gym, Atari Learning
Environment, or Malmo which allow agents to learn complex tasks through
interaction with virtual environments. While RL is also increasingly applied to
natural language processing (NLP), there are no simulated textual environments
available for researchers to apply and consistently benchmark RL on NLP tasks.
With the work reported here, we therefore release NLPGym, an open-source Python
toolkit that provides interactive textual environments for standard NLP tasks
such as sequence tagging, multi-label classification, and question answering.
We also present experimental results for 6 tasks using different RL algorithms
which serve as baselines for further research. The toolkit is published at
https://github.com/rajcscw/nlp-gym
Related papers
- Exploring RL-based LLM Training for Formal Language Tasks with Programmed Rewards [49.7719149179179]
This paper investigates the feasibility of using PPO for reinforcement learning (RL) from explicitly programmed reward signals.
We focus on tasks expressed through formal languages, such as programming, where explicit reward functions can be programmed to automatically assess quality of generated outputs.
Our results show that pure RL-based training for the two formal language tasks is challenging, with success being limited even for the simple arithmetic task.
arXiv Detail & Related papers (2024-10-22T15:59:58Z) - Bridging Environments and Language with Rendering Functions and Vision-Language Models [7.704773649029078]
Vision-language models (VLMs) have tremendous potential for grounding language.
This paper introduces a novel decomposition of the problem of building language-conditioned agents (LCAs)
We also explore several enhancements to the speed and quality of VLM-based LCAs.
arXiv Detail & Related papers (2024-09-24T12:24:07Z) - YAMLE: Yet Another Machine Learning Environment [4.985768723667417]
YAMLE is an open-source framework that facilitates rapid prototyping and experimentation with machine learning (ML) models and methods.
YAMLE includes a command-line interface and integrations with popular and well-maintained PyTorch-based libraries.
The ambition for YAMLE is to grow into a shared ecosystem where researchers and practitioners can quickly build on and compare existing implementations.
arXiv Detail & Related papers (2024-02-09T09:34:36Z) - Interactive Planning Using Large Language Models for Partially
Observable Robotics Tasks [54.60571399091711]
Large Language Models (LLMs) have achieved impressive results in creating robotic agents for performing open vocabulary tasks.
We present an interactive planning technique for partially observable tasks using LLMs.
arXiv Detail & Related papers (2023-12-11T22:54:44Z) - Is Reinforcement Learning (Not) for Natural Language Processing?:
Benchmarks, Baselines, and Building Blocks for Natural Language Policy
Optimization [73.74371798168642]
We introduce an open-source modular library, RL4LMs, for optimizing language generators with reinforcement learning.
Next, we present the GRUE benchmark, a set of 6 language generation tasks which are supervised not by target strings, but by reward functions.
Finally, we introduce an easy-to-use, performant RL algorithm, NLPO, that learns to effectively reduce the action space in language generation.
arXiv Detail & Related papers (2022-10-03T21:38:29Z) - pymdp: A Python library for active inference in discrete state spaces [52.85819390191516]
pymdp is an open-source package for simulating active inference in Python.
We provide the first open-source package for simulating active inference with POMDPs.
arXiv Detail & Related papers (2022-01-11T12:18:44Z) - Modular Procedural Generation for Voxel Maps [2.6811189633660613]
In this paper, we present mcg, an open-source library to facilitate implementing PCG algorithms for voxel-based environments such as Minecraft.
The library is designed with human-machine teaming research in mind, and thus takes a 'top-down' approach to generation.
The benefits of this approach include rapid, scalable, and efficient development of virtual environments, the ability to control the statistics of the environment at a semantic level, and the ability to generate novel environments in response to player actions in real time.
arXiv Detail & Related papers (2021-04-18T16:21:35Z) - TeachMyAgent: a Benchmark for Automatic Curriculum Learning in Deep RL [23.719833581321033]
Training autonomous agents able to generalize to multiple tasks is a key target of Deep Reinforcement Learning (DRL) research.
In parallel to improving DRL algorithms, Automatic Curriculum Learning (ACL) study how teacher algorithms can train DRL agents more efficiently by adapting task selection to their evolving abilities.
While multiple standard benchmarks exist to compare DRL agents, there is currently no such thing for ACL algorithms.
arXiv Detail & Related papers (2021-03-17T17:59:22Z) - Discovering Reinforcement Learning Algorithms [53.72358280495428]
Reinforcement learning algorithms update an agent's parameters according to one of several possible rules.
This paper introduces a new meta-learning approach that discovers an entire update rule.
It includes both 'what to predict' (e.g. value functions) and 'how to learn from it' by interacting with a set of environments.
arXiv Detail & Related papers (2020-07-17T07:38:39Z) - The NetHack Learning Environment [79.06395964379107]
We present the NetHack Learning Environment (NLE), a procedurally generated rogue-like environment for Reinforcement Learning research.
We argue that NetHack is sufficiently complex to drive long-term research on problems such as exploration, planning, skill acquisition, and language-conditioned RL.
We demonstrate empirical success for early stages of the game using a distributed Deep RL baseline and Random Network Distillation exploration.
arXiv Detail & Related papers (2020-06-24T14:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.