Action Space Shaping in Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2004.00980v2
- Date: Tue, 26 May 2020 09:25:59 GMT
- Title: Action Space Shaping in Deep Reinforcement Learning
- Authors: Anssi Kanervisto, Christian Scheller, Ville Hautam\"aki
- Abstract summary: Reinforcement learning has been successful in training agents in various learning environments, including video-games.
We aim to gain insight on these action space modifications by conducting extensive experiments in video-game environments.
Our results show how domain-specific removal of actions and discretization of continuous actions can be crucial for successful learning.
- Score: 7.508516104014916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL) has been successful in training agents in various
learning environments, including video-games. However, such work modifies and
shrinks the action space from the game's original. This is to avoid trying
"pointless" actions and to ease the implementation. Currently, this is mostly
done based on intuition, with little systematic research supporting the design
decisions. In this work, we aim to gain insight on these action space
modifications by conducting extensive experiments in video-game environments.
Our results show how domain-specific removal of actions and discretization of
continuous actions can be crucial for successful learning. With these insights,
we hope to ease the use of RL in new environments, by clarifying what
action-spaces are easy to learn.
Related papers
- Latent Action Priors From a Single Gait Cycle Demonstration for Online Imitation Learning [42.642008092347986]
We propose an additional inductive bias for robot learning: latent actions learned from expert demonstration as priors in the action space.
We show that these action priors can be learned from only a single open-loop gait cycle using a simple autoencoder.
arXiv Detail & Related papers (2024-10-04T09:10:56Z) - Can VLMs Play Action Role-Playing Games? Take Black Myth Wukong as a Study Case [20.14197375326218]
This research aims to provide new insights and directions for applying multimodal agents in complex action game environments.
We select an ARPG, Black Myth: Wukong'', as a research platform to explore the capability boundaries of existing vision language models.
We will release a human operation dataset containing recorded gameplay videos and operation logs, including mouse and keyboard actions.
arXiv Detail & Related papers (2024-09-19T16:30:25Z) - A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free
Reinforcement Learning [86.06110576808824]
Deep reinforcement learning is a promising approach to learning policies in uncontrolled environments.
Recent advancements in machine learning algorithms and libraries combined with a carefully tuned robot controller lead to learning quadruped in only 20 minutes in the real world.
arXiv Detail & Related papers (2022-08-16T17:37:36Z) - Basis for Intentions: Efficient Inverse Reinforcement Learning using
Past Experience [89.30876995059168]
inverse reinforcement learning (IRL) -- inferring the reward function of an agent from observing its behavior.
This paper addresses the problem of IRL -- inferring the reward function of an agent from observing its behavior.
arXiv Detail & Related papers (2022-08-09T17:29:49Z) - A Survey on Reinforcement Learning Methods in Character Animation [22.3342752080749]
Reinforcement Learning is an area of Machine Learning focused on how agents can be trained to make sequential decisions.
This paper surveys the modern Deep Reinforcement Learning methods and discusses their possible applications in Character Animation.
arXiv Detail & Related papers (2022-03-07T23:39:00Z) - TRAIL: Near-Optimal Imitation Learning with Suboptimal Data [100.83688818427915]
We present training objectives that use offline datasets to learn a factored transition model.
Our theoretical analysis shows that the learned latent action space can boost the sample-efficiency of downstream imitation learning.
To learn the latent action space in practice, we propose TRAIL (Transition-Reparametrized Actions for Imitation Learning), an algorithm that learns an energy-based transition model.
arXiv Detail & Related papers (2021-10-27T21:05:00Z) - LASER: Learning a Latent Action Space for Efficient Reinforcement
Learning [41.53297694894669]
We present LASER, a method to learn latent action spaces for efficient reinforcement learning.
We show improved sample efficiency compared to the original action space from better alignment of the action space to the task space, as we observe with visualizations of the learned action space manifold.
arXiv Detail & Related papers (2021-03-29T17:40:02Z) - Disturbing Reinforcement Learning Agents with Corrupted Rewards [62.997667081978825]
We analyze the effects of different attack strategies based on reward perturbations on reinforcement learning algorithms.
We show that smoothly crafting adversarial rewards are able to mislead the learner, and that using low exploration probability values, the policy learned is more robust to corrupt rewards.
arXiv Detail & Related papers (2021-02-12T15:53:48Z) - Demonstration-efficient Inverse Reinforcement Learning in Procedurally
Generated Environments [137.86426963572214]
Inverse Reinforcement Learning can extrapolate reward functions from expert demonstrations.
We show that our approach, DE-AIRL, is demonstration-efficient and still able to extrapolate reward functions which generalize to the fully procedural domain.
arXiv Detail & Related papers (2020-12-04T11:18:02Z) - Exploration Based Language Learning for Text-Based Games [72.30525050367216]
This work presents an exploration and imitation-learning-based agent capable of state-of-the-art performance in playing text-based computer games.
Text-based computer games describe their world to the player through natural language and expect the player to interact with the game using text.
These games are of interest as they can be seen as a testbed for language understanding, problem-solving, and language generation by artificial agents.
arXiv Detail & Related papers (2020-01-24T03:03:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.