Deep Latent Competition: Learning to Race Using Visual Control Policies
in Latent Space
- URL: http://arxiv.org/abs/2102.09812v1
- Date: Fri, 19 Feb 2021 09:00:29 GMT
- Title: Deep Latent Competition: Learning to Race Using Visual Control Policies
in Latent Space
- Authors: Wilko Schwarting, Tim Seyde, Igor Gilitschenski, Lucas Liebenwein,
Ryan Sander, Sertac Karaman, Daniela Rus
- Abstract summary: Deep Latent Competition (DLC) is a reinforcement learning algorithm that learns competitive visual control policies through self-play in imagination.
Imagined self-play reduces costly sample generation in the real world, while the latent representation enables planning to scale gracefully with observation dimensionality.
- Score: 63.57289340402389
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning competitive behaviors in multi-agent settings such as racing
requires long-term reasoning about potential adversarial interactions. This
paper presents Deep Latent Competition (DLC), a novel reinforcement learning
algorithm that learns competitive visual control policies through self-play in
imagination. The DLC agent imagines multi-agent interaction sequences in the
compact latent space of a learned world model that combines a joint transition
function with opponent viewpoint prediction. Imagined self-play reduces costly
sample generation in the real world, while the latent representation enables
planning to scale gracefully with observation dimensionality. We demonstrate
the effectiveness of our algorithm in learning competitive behaviors on a novel
multi-agent racing benchmark that requires planning from image observations.
Code and videos available at
https://sites.google.com/view/deep-latent-competition.
Related papers
- Learn 2 Rage: Experiencing The Emotional Roller Coaster That Is Reinforcement Learning [5.962453678471195]
This work presents the experiments and solution outline for our teams winning submission in the Learn To Race Autonomous Racing Virtual Challenge 2022 hosted by AIcrowd.
The objective of the Learn-to-Race competition is to push the boundary of autonomous technology, with a focus on achieving the safety benefits of autonomous driving.
We focused our initial efforts on implementation of Soft Actor Critic (SAC) variants.
Our goal was to learn non-trivial control of the race car exclusively from visual and geometric features, directly mapping pixels to control actions.
arXiv Detail & Related papers (2024-10-24T06:16:52Z) - Learning Robot Soccer from Egocentric Vision with Deep Reinforcement Learning [17.906144781244336]
We train end-to-end robot soccer policies with fully onboard computation and sensing via egocentric RGB vision.
This paper constitutes a first demonstration of end-to-end training for multi-agent robot soccer.
arXiv Detail & Related papers (2024-05-03T18:41:13Z) - Learning of Generalizable and Interpretable Knowledge in Grid-Based
Reinforcement Learning Environments [5.217870815854702]
We propose using program synthesis to imitate reinforcement learning policies.
We adapt the state-of-the-art program synthesis system DreamCoder for learning concepts in grid-based environments.
arXiv Detail & Related papers (2023-09-07T11:46:57Z) - Finding mixed-strategy equilibria of continuous-action games without
gradients using randomized policy networks [83.28949556413717]
We study the problem of computing an approximate Nash equilibrium of continuous-action game without access to gradients.
We model players' strategies using artificial neural networks.
This paper is the first to solve general continuous-action games with unrestricted mixed strategies and without any gradient information.
arXiv Detail & Related papers (2022-11-29T05:16:41Z) - Stochastic Coherence Over Attention Trajectory For Continuous Learning
In Video Streams [64.82800502603138]
This paper proposes a novel neural-network-based approach to progressively and autonomously develop pixel-wise representations in a video stream.
The proposed method is based on a human-like attention mechanism that allows the agent to learn by observing what is moving in the attended locations.
Our experiments leverage 3D virtual environments and they show that the proposed agents can learn to distinguish objects just by observing the video stream.
arXiv Detail & Related papers (2022-04-26T09:52:31Z) - Time-series Imputation of Temporally-occluded Multiagent Trajectories [18.862173210927658]
We study the problem of multiagent time-series imputation, where available past and future observations of subsets of agents are used to estimate missing observations for other agents.
Our approach, called the Graph Imputer, uses forward- and backward-information in combination with graph networks and variational autoencoders.
We evaluate our approach on a dataset of football matches, using a projective camera module to train and evaluate our model for the off-screen player state estimation setting.
arXiv Detail & Related papers (2021-06-08T09:58:43Z) - CoCon: Cooperative-Contrastive Learning [52.342936645996765]
Self-supervised visual representation learning is key for efficient video analysis.
Recent success in learning image representations suggests contrastive learning is a promising framework to tackle this challenge.
We introduce a cooperative variant of contrastive learning to utilize complementary information across views.
arXiv Detail & Related papers (2021-04-30T05:46:02Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z) - Learning from Learners: Adapting Reinforcement Learning Agents to be
Competitive in a Card Game [71.24825724518847]
We present a study on how popular reinforcement learning algorithms can be adapted to learn and to play a real-world implementation of a competitive multiplayer card game.
We propose specific training and validation routines for the learning agents, in order to evaluate how the agents learn to be competitive and explain how they adapt to each others' playing style.
arXiv Detail & Related papers (2020-04-08T14:11:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.