RARA: Zero-shot Sim2Real Visual Navigation with Following Foreground
Cues
- URL: http://arxiv.org/abs/2201.02798v1
- Date: Sat, 8 Jan 2022 09:53:21 GMT
- Title: RARA: Zero-shot Sim2Real Visual Navigation with Following Foreground
Cues
- Authors: Klaas Kelchtermans, Tinne Tuytelaars
- Abstract summary: We tackle the specific case of camera-based navigation, formulating it as following a visual cue in the foreground with arbitrary backgrounds.
The goal is to train a visual agent on data captured in an empty simulated environment except for this foreground cue and test this model directly in a visually diverse real world.
- Score: 42.998649025215045
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The gap between simulation and the real-world restrains many machine learning
breakthroughs in computer vision and reinforcement learning from being
applicable in the real world. In this work, we tackle this gap for the specific
case of camera-based navigation, formulating it as following a visual cue in
the foreground with arbitrary backgrounds. The visual cue in the foreground can
often be simulated realistically, such as a line, gate or cone. The challenge
then lies in coping with the unknown backgrounds and integrating both. As such,
the goal is to train a visual agent on data captured in an empty simulated
environment except for this foreground cue and test this model directly in a
visually diverse real world. In order to bridge this big gap, we show it's
crucial to combine following techniques namely: Randomized augmentation of the
fore- and background, regularization with both deep supervision and triplet
loss and finally abstraction of the dynamics by using waypoints rather than
direct velocity commands. The various techniques are ablated in our
experimental results both qualitatively and quantitatively finally
demonstrating a successful transfer from simulation to the real world.
Related papers
- Gaussian Splatting to Real World Flight Navigation Transfer with Liquid Networks [93.38375271826202]
We present a method to improve generalization and robustness to distribution shifts in sim-to-real visual quadrotor navigation tasks.
We first build a simulator by integrating Gaussian splatting with quadrotor flight dynamics, and then, train robust navigation policies using Liquid neural networks.
In this way, we obtain a full-stack imitation learning protocol that combines advances in 3D Gaussian splatting radiance field rendering, programming of expert demonstration training data, and the task understanding capabilities of Liquid networks.
arXiv Detail & Related papers (2024-06-21T13:48:37Z) - Scaling Face Interaction Graph Networks to Real World Scenes [12.519862235430153]
We introduce a method which substantially reduces the memory required to run graph-based learned simulators.
We show that our method uses substantially less memory than previous graph-based simulators while retaining their accuracy.
This paves the way for expanding the application of learned simulators to settings where only perceptual information is available at inference time.
arXiv Detail & Related papers (2024-01-22T14:38:25Z) - Sim-to-Real via Sim-to-Seg: End-to-end Off-road Autonomous Driving
Without Real Data [56.49494318285391]
We present Sim2Seg, a re-imagining of RCAN that crosses the visual reality gap for off-road autonomous driving.
This is done by learning to translate randomized simulation images into simulated segmentation and depth maps.
This allows us to train an end-to-end RL policy in simulation, and directly deploy in the real-world.
arXiv Detail & Related papers (2022-10-25T17:50:36Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z) - Practical Imitation Learning in the Real World via Task Consistency Loss [18.827979446629296]
This paper introduces a self-supervised loss that encourages sim and real alignment both at the feature and action-prediction levels.
We achieve 80% success across ten seen and unseen scenes using only 16.2 hours of teleoperated demonstrations in sim and real.
arXiv Detail & Related papers (2022-02-03T21:43:06Z) - Towards Optimal Strategies for Training Self-Driving Perception Models
in Simulation [98.51313127382937]
We focus on the use of labels in the synthetic domain alone.
Our approach introduces both a way to learn neural-invariant representations and a theoretically inspired view on how to sample the data from the simulator.
We showcase our approach on the bird's-eye-view vehicle segmentation task with multi-sensor data.
arXiv Detail & Related papers (2021-11-15T18:37:43Z) - Vision-Guided Quadrupedal Locomotion in the Wild with Multi-Modal Delay
Randomization [9.014518402531875]
We train the RL policy for end-to-end control in a physical simulator without any predefined controller or reference motion.
We demonstrate the robot can smoothly maneuver at a high speed, avoid the obstacles, and show significant improvement over the baselines.
arXiv Detail & Related papers (2021-09-29T16:48:05Z) - SimAug: Learning Robust Representations from Simulation for Trajectory
Prediction [78.91518036949918]
We propose a novel approach to learn robust representation through augmenting the simulation training data.
We show that SimAug achieves promising results on three real-world benchmarks using zero real training data.
arXiv Detail & Related papers (2020-04-04T21:22:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.