Generalization of Agent Behavior through Explicit Representation of
Context
- URL: http://arxiv.org/abs/2006.11305v2
- Date: Tue, 9 Feb 2021 21:51:41 GMT
- Title: Generalization of Agent Behavior through Explicit Representation of
Context
- Authors: Cem C Tutum, Suhaib Abdulquddos, Risto Miikkulainen
- Abstract summary: In order to deploy autonomous agents in digital interactive environments, they must be able to act robustly in unseen situations.
This paper proposes a principled approach where a context module is coevolved with a skill module in the game.
The approach is evaluated in the Flappy Bird and LunarLander video games, as well as in the CARLA autonomous driving simulation.
- Score: 14.272883554753323
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In order to deploy autonomous agents in digital interactive environments,
they must be able to act robustly in unseen situations. The standard machine
learning approach is to include as much variation as possible into training
these agents. The agents can then interpolate within their training, but they
cannot extrapolate much beyond it. This paper proposes a principled approach
where a context module is coevolved with a skill module in the game. The
context module recognizes the temporal variation in the game and modulates the
outputs of the skill module so that the action decisions can be made robustly
even in previously unseen situations. The approach is evaluated in the Flappy
Bird and LunarLander video games, as well as in the CARLA autonomous driving
simulation. The Context+Skill approach leads to significantly more robust
behavior in environments that require extrapolation beyond training. Such a
principled generalization ability is essential in deploying autonomous agents
in real-world tasks, and can serve as a foundation for continual adaptation as
well.
Related papers
- Aligning Agents like Large Language Models [8.873319874424167]
Training agents to behave as desired in complex 3D environments from high-dimensional sensory information is challenging.
We draw an analogy between the undesirable behaviors of imitation learning agents and the unhelpful responses of unaligned large language models (LLMs)
We demonstrate that we can align our agent to consistently perform the desired mode, while providing insights and advice for successfully applying this approach to training agents.
arXiv Detail & Related papers (2024-06-06T16:05:45Z) - HAZARD Challenge: Embodied Decision Making in Dynamically Changing
Environments [93.94020724735199]
HAZARD consists of three unexpected disaster scenarios, including fire, flood, and wind.
This benchmark enables us to evaluate autonomous agents' decision-making capabilities across various pipelines.
arXiv Detail & Related papers (2024-01-23T18:59:43Z) - Pangu-Agent: A Fine-Tunable Generalist Agent with Structured Reasoning [50.47568731994238]
Key method for creating Artificial Intelligence (AI) agents is Reinforcement Learning (RL)
This paper presents a general framework model for integrating and learning structured reasoning into AI agents' policies.
arXiv Detail & Related papers (2023-12-22T17:57:57Z) - Interactive Autonomous Navigation with Internal State Inference and
Interactivity Estimation [58.21683603243387]
We propose three auxiliary tasks with relational-temporal reasoning and integrate them into the standard Deep Learning framework.
These auxiliary tasks provide additional supervision signals to infer the behavior patterns other interactive agents.
Our approach achieves robust and state-of-the-art performance in terms of standard evaluation metrics.
arXiv Detail & Related papers (2023-11-27T18:57:42Z) - Drive Anywhere: Generalizable End-to-end Autonomous Driving with
Multi-modal Foundation Models [114.69732301904419]
We present an approach to apply end-to-end open-set (any environment/scene) autonomous driving that is capable of providing driving decisions from representations queryable by image and text.
Our approach demonstrates unparalleled results in diverse tests while achieving significantly greater robustness in out-of-distribution situations.
arXiv Detail & Related papers (2023-10-26T17:56:35Z) - Hierarchical Imitation Learning for Stochastic Environments [31.64016324441371]
Existing methods that improve distributional realism typically rely on hierarchical policies.
We propose Robust Type Conditioning (RTC), which eliminates the shift with adversarial training under environmentality.
Experiments on two domains, including the large-scale Open Motion dataset, show improved distributional realism while maintaining or improving task performance compared to state-of-the-art baselines.
arXiv Detail & Related papers (2023-09-25T10:10:34Z) - A Survey on Reinforcement Learning Methods in Character Animation [22.3342752080749]
Reinforcement Learning is an area of Machine Learning focused on how agents can be trained to make sequential decisions.
This paper surveys the modern Deep Reinforcement Learning methods and discusses their possible applications in Character Animation.
arXiv Detail & Related papers (2022-03-07T23:39:00Z) - TrafficSim: Learning to Simulate Realistic Multi-Agent Behaviors [74.67698916175614]
We propose TrafficSim, a multi-agent behavior model for realistic traffic simulation.
In particular, we leverage an implicit latent variable model to parameterize a joint actor policy.
We show TrafficSim generates significantly more realistic and diverse traffic scenarios as compared to a diverse set of baselines.
arXiv Detail & Related papers (2021-01-17T00:29:30Z) - Adapting to Unseen Environments through Explicit Representation of
Context [16.8615211682877]
In order to deploy autonomous agents to domains such as autonomous driving, infrastructure management, health care, and finance, they must be able to adapt safely to unseen situations.
This paper proposes a principled approach where a context module is coevolved with a skill module.
The Context+Skill approach leads to significantly more robust behavior in environments with previously unseen effects.
arXiv Detail & Related papers (2020-02-13T17:15:47Z) - Intelligent Roundabout Insertion using Deep Reinforcement Learning [68.8204255655161]
We present a maneuver planning module able to negotiate the entering in busy roundabouts.
The proposed module is based on a neural network trained to predict when and how entering the roundabout throughout the whole duration of the maneuver.
arXiv Detail & Related papers (2020-01-03T11:16:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.