Intuitive Physics Guided Exploration for Sample Efficient Sim2real
Transfer
- URL: http://arxiv.org/abs/2104.08795v1
- Date: Sun, 18 Apr 2021 10:03:26 GMT
- Title: Intuitive Physics Guided Exploration for Sample Efficient Sim2real
Transfer
- Authors: Buddhika Laknath Semage, Thommen George Karimpanal, Santu Rana, Svetha
Venkatesh
- Abstract summary: This paper focuses on learning task-specific estimates of latent factors which allow approximation of real world trajectories in an ideal simulation environment.
We first introduce intuitive action groupings based on human physics knowledge and experience, which is then used to design novel strategies for interacting with the real environment.
We demonstrate our approach in a range of physics based tasks, and show that it achieves superior performance relative to other baselines, using only a limited number of real-world interactions.
- Score: 42.23861067181556
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Physics-based reinforcement learning tasks can benefit from simplified
physics simulators as they potentially allow near-optimal policies to be
learned in simulation. However, such simulators require the latent factors
(e.g. mass, friction coefficient etc.) of the associated objects and other
environment-specific factors (e.g. wind speed, air density etc.) to be
accurately specified, without which, it could take considerable additional
learning effort to adapt the learned simulation policy to the real environment.
As such a complete specification can be impractical, in this paper, we instead,
focus on learning task-specific estimates of latent factors which allow the
approximation of real world trajectories in an ideal simulation environment.
Specifically, we propose two new concepts: a) action grouping - the idea that
certain types of actions are closely associated with the estimation of certain
latent factors, and; b) partial grounding - the idea that simulation of
task-specific dynamics may not need precise estimation of all the latent
factors. We first introduce intuitive action groupings based on human physics
knowledge and experience, which is then used to design novel strategies for
interacting with the real environment. Next, we describe how prior knowledge of
a task in a given environment can be used to extract the relative importance of
different latent factors, and how this can be used to inform partial grounding,
which enables efficient learning of the task in any arbitrary environment. We
demonstrate our approach in a range of physics based tasks, and show that it
achieves superior performance relative to other baselines, using only a limited
number of real-world interactions.
Related papers
- Reward Function Design for Crowd Simulation via Reinforcement Learning [12.449513548800466]
Reinforcement learning has shown great potential in simulating virtual crowds, but the design of the reward function is critical to achieving effective and efficient results.
We provide theoretical insights on the validity of certain reward functions according to their analytical properties, and evaluate them empirically using a range of scenarios.
Our findings can inform the development of new crowd simulation techniques, and contribute to the wider study of human-like navigation.
arXiv Detail & Related papers (2023-09-22T12:55:30Z) - AdaptSim: Task-Driven Simulation Adaptation for Sim-to-Real Transfer [10.173835871228718]
AdaptSim aims to optimize task performance in target (real) environments.
First, we meta-learn an adaptation policy in simulation using reinforcement learning.
We then perform iterative real-world adaptation by inferring new simulation parameter distributions for policy training.
arXiv Detail & Related papers (2023-02-09T19:10:57Z) - Towards Autonomous Grading In The Real World [4.651327752886103]
We aim to tackle the problem of autonomous grading, where a dozer is required to flatten an uneven area.
We design both a realistic physical simulation and a scaled real prototype environment mimicking the real dozer dynamics and sensory information.
arXiv Detail & Related papers (2022-06-13T12:21:20Z) - An in-depth experimental study of sensor usage and visual reasoning of
robots navigating in real environments [20.105395754497202]
We study the performance and reasoning capacities of real physical agents, trained in simulation and deployed to two different physical environments.
We show, that for the PointGoal task, an agent pre-trained on wide variety of tasks and fine-tuned on a simulated version of the target environment can reach competitive performance without modelling any sim2real transfer.
arXiv Detail & Related papers (2021-11-29T16:27:29Z) - BEHAVIOR: Benchmark for Everyday Household Activities in Virtual,
Interactive, and Ecological Environments [70.18430114842094]
We introduce BEHAVIOR, a benchmark for embodied AI with 100 activities in simulation.
These activities are designed to be realistic, diverse, and complex.
We include 500 human demonstrations in virtual reality (VR) to serve as the human ground truth.
arXiv Detail & Related papers (2021-08-06T23:36:23Z) - Zero-Shot Reinforcement Learning on Graphs for Autonomous Exploration
Under Uncertainty [6.42522897323111]
We present a framework for self-learning a high-performance exploration policy in a single simulation environment.
We propose a novel approach that uses graph neural networks in conjunction with deep reinforcement learning.
arXiv Detail & Related papers (2021-05-11T02:42:17Z) - TrafficSim: Learning to Simulate Realistic Multi-Agent Behaviors [74.67698916175614]
We propose TrafficSim, a multi-agent behavior model for realistic traffic simulation.
In particular, we leverage an implicit latent variable model to parameterize a joint actor policy.
We show TrafficSim generates significantly more realistic and diverse traffic scenarios as compared to a diverse set of baselines.
arXiv Detail & Related papers (2021-01-17T00:29:30Z) - Reactive Long Horizon Task Execution via Visual Skill and Precondition
Models [59.76233967614774]
We describe an approach for sim-to-real training that can accomplish unseen robotic tasks using models learned in simulation to ground components of a simple task planner.
We show an increase in success rate from 91.6% to 98% in simulation and from 10% to 80% success rate in the real-world as compared with naive baselines.
arXiv Detail & Related papers (2020-11-17T15:24:01Z) - CausalWorld: A Robotic Manipulation Benchmark for Causal Structure and
Transfer Learning [138.40338621974954]
CausalWorld is a benchmark for causal structure and transfer learning in a robotic manipulation environment.
Tasks consist of constructing 3D shapes from a given set of blocks - inspired by how children learn to build complex structures.
arXiv Detail & Related papers (2020-10-08T23:01:13Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.