Learning a State Representation and Navigation in Cluttered and Dynamic
Environments
- URL: http://arxiv.org/abs/2103.04351v1
- Date: Sun, 7 Mar 2021 13:19:06 GMT
- Title: Learning a State Representation and Navigation in Cluttered and Dynamic
Environments
- Authors: David Hoeller, Lorenz Wellhausen, Farbod Farshidian, Marco Hutter
- Abstract summary: We present a learning-based pipeline to realise local navigation with a quadrupedal robot in cluttered environments.
The robot is able to safely locomote to a target location based on frames from a depth camera without any explicit mapping of the environment.
We show that our system can handle noisy depth images, avoid dynamic obstacles unseen during training, and is endowed with local spatial awareness.
- Score: 6.909283975004628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we present a learning-based pipeline to realise local
navigation with a quadrupedal robot in cluttered environments with static and
dynamic obstacles. Given high-level navigation commands, the robot is able to
safely locomote to a target location based on frames from a depth camera
without any explicit mapping of the environment. First, the sequence of images
and the current trajectory of the camera are fused to form a model of the world
using state representation learning. The output of this lightweight module is
then directly fed into a target-reaching and obstacle-avoiding policy trained
with reinforcement learning. We show that decoupling the pipeline into these
components results in a sample efficient policy learning stage that can be
fully trained in simulation in just a dozen minutes. The key part is the state
representation, which is trained to not only estimate the hidden state of the
world in an unsupervised fashion, but also helps bridging the reality gap,
enabling successful sim-to-real transfer. In our experiments with the
quadrupedal robot ANYmal in simulation and in reality, we show that our system
can handle noisy depth images, avoid dynamic obstacles unseen during training,
and is endowed with local spatial awareness.
Related papers
- Flow as the Cross-Domain Manipulation Interface [73.15952395641136]
Im2Flow2Act enables robots to acquire real-world manipulation skills without the need of real-world robot training data.
Im2Flow2Act comprises two components: a flow generation network and a flow-conditioned policy.
We demonstrate Im2Flow2Act's capabilities in a variety of real-world tasks, including the manipulation of rigid, articulated, and deformable objects.
arXiv Detail & Related papers (2024-07-21T16:15:02Z) - Gaussian Splatting to Real World Flight Navigation Transfer with Liquid Networks [93.38375271826202]
We present a method to improve generalization and robustness to distribution shifts in sim-to-real visual quadrotor navigation tasks.
We first build a simulator by integrating Gaussian splatting with quadrotor flight dynamics, and then, train robust navigation policies using Liquid neural networks.
In this way, we obtain a full-stack imitation learning protocol that combines advances in 3D Gaussian splatting radiance field rendering, programming of expert demonstration training data, and the task understanding capabilities of Liquid networks.
arXiv Detail & Related papers (2024-06-21T13:48:37Z) - Navigating to Objects in the Real World [76.1517654037993]
We present a large-scale empirical study of semantic visual navigation methods comparing methods from classical, modular, and end-to-end learning approaches.
We find that modular learning works well in the real world, attaining a 90% success rate.
In contrast, end-to-end learning does not, dropping from 77% simulation to 23% real-world success rate due to a large image domain gap between simulation and reality.
arXiv Detail & Related papers (2022-12-02T01:10:47Z) - On the Application of Efficient Neural Mapping to Real-Time Indoor
Localisation for Unmanned Ground Vehicles [5.137284292672375]
We show that a compact model capable of real-time inference on embedded platforms can be used to achieve localisation accuracy of several centimetres.
We deploy our trained model onboard a UGV platform, demonstrating its effectiveness in a waypoint navigation task.
arXiv Detail & Related papers (2022-11-09T07:23:28Z) - Sim-to-Real via Sim-to-Seg: End-to-end Off-road Autonomous Driving
Without Real Data [56.49494318285391]
We present Sim2Seg, a re-imagining of RCAN that crosses the visual reality gap for off-road autonomous driving.
This is done by learning to translate randomized simulation images into simulated segmentation and depth maps.
This allows us to train an end-to-end RL policy in simulation, and directly deploy in the real-world.
arXiv Detail & Related papers (2022-10-25T17:50:36Z) - Neural Scene Representation for Locomotion on Structured Terrain [56.48607865960868]
We propose a learning-based method to reconstruct the local terrain for a mobile robot traversing urban environments.
Using a stream of depth measurements from the onboard cameras and the robot's trajectory, the estimates the topography in the robot's vicinity.
We propose a 3D reconstruction model that faithfully reconstructs the scene, despite the noisy measurements and large amounts of missing data coming from the blind spots of the camera arrangement.
arXiv Detail & Related papers (2022-06-16T10:45:17Z) - On Embodied Visual Navigation in Real Environments Through Habitat [20.630139085937586]
Visual navigation models based on deep learning can learn effective policies when trained on large amounts of visual observations.
To deal with this limitation, several simulation platforms have been proposed in order to train visual navigation policies on virtual environments efficiently.
We show that our tool can effectively help to train and evaluate navigation policies on real-world observations without running navigation pisodes in the real world.
arXiv Detail & Related papers (2020-10-26T09:19:07Z) - Embodied Visual Navigation with Automatic Curriculum Learning in Real
Environments [20.017277077448924]
NavACL is a method of automatic curriculum learning tailored to the navigation task.
Deep reinforcement learning agents trained using NavACL significantly outperform state-of-the-art agents trained with uniform sampling.
Our agents can navigate through unknown cluttered indoor environments to semantically-specified targets using only RGB images.
arXiv Detail & Related papers (2020-09-11T13:28:26Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z) - Stillleben: Realistic Scene Synthesis for Deep Learning in Robotics [33.30312206728974]
We describe a synthesis pipeline capable of producing training data for cluttered scene perception tasks.
Our approach arranges object meshes in physically realistic, dense scenes using physics simulation.
Our pipeline can be run online during training of a deep neural network.
arXiv Detail & Related papers (2020-05-12T10:11:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.