Learning to Grasp on the Moon from 3D Octree Observations with Deep
Reinforcement Learning
- URL: http://arxiv.org/abs/2208.00818v1
- Date: Mon, 1 Aug 2022 12:59:03 GMT
- Title: Learning to Grasp on the Moon from 3D Octree Observations with Deep
Reinforcement Learning
- Authors: Andrej Orsula, Simon B{\o}gh, Miguel Olivares-Mendez and Carol
Martinez
- Abstract summary: This work investigates the applicability of deep reinforcement learning for vision-based robotic grasping of objects on the Moon.
A novel simulation environment with procedurally-generated datasets is created to train agents under challenging conditions.
A model-free off-policy actor-critic algorithm is then employed for end-to-end learning of a policy.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Extraterrestrial rovers with a general-purpose robotic arm have many
potential applications in lunar and planetary exploration. Introducing autonomy
into such systems is desirable for increasing the time that rovers can spend
gathering scientific data and collecting samples. This work investigates the
applicability of deep reinforcement learning for vision-based robotic grasping
of objects on the Moon. A novel simulation environment with
procedurally-generated datasets is created to train agents under challenging
conditions in unstructured scenes with uneven terrain and harsh illumination. A
model-free off-policy actor-critic algorithm is then employed for end-to-end
learning of a policy that directly maps compact octree observations to
continuous actions in Cartesian space. Experimental evaluation indicates that
3D data representations enable more effective learning of manipulation skills
when compared to traditionally used image-based observations. Domain
randomization improves the generalization of learned policies to novel scenes
with previously unseen objects and different illumination conditions. To this
end, we demonstrate zero-shot sim-to-real transfer by evaluating trained agents
on a real robot in a Moon-analogue facility.
Related papers
- ManiBox: Enhancing Spatial Grasping Generalization via Scalable Simulation Data Generation [37.73074657448699]
bfManiBox is a novel bounding-box-guided manipulation method built on a simulation-based teacher-student framework.
ManiBox demonstrates a marked improvement in spatial grasping generalization and adaptability to diverse objects and backgrounds.
arXiv Detail & Related papers (2024-11-04T07:05:02Z) - Learning autonomous driving from aerial imagery [67.06858775696453]
Photogrammetric simulators allow the synthesis of novel views through the transformation of pre-generated assets into novel views.
We use a Neural Radiance Field (NeRF) as an intermediate representation to synthesize novel views from the point of view of a ground vehicle.
arXiv Detail & Related papers (2024-10-18T05:09:07Z) - Reconstructing Objects in-the-wild for Realistic Sensor Simulation [41.55571880832957]
We present NeuSim, a novel approach that estimates accurate geometry and realistic appearance from sparse in-the-wild data.
We model the object appearance with a robust physics-inspired reflectance representation effective for in-the-wild data.
Our experiments show that NeuSim has strong view synthesis performance on challenging scenarios with sparse training views.
arXiv Detail & Related papers (2023-11-09T18:58:22Z) - 3D-IntPhys: Towards More Generalized 3D-grounded Visual Intuitive
Physics under Challenging Scenes [68.66237114509264]
We present a framework capable of learning 3D-grounded visual intuitive physics models from videos of complex scenes with fluids.
We show our model can make long-horizon future predictions by learning from raw images and significantly outperforms models that do not employ an explicit 3D representation space.
arXiv Detail & Related papers (2023-04-22T19:28:49Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - RandomRooms: Unsupervised Pre-training from Synthetic Shapes and
Randomized Layouts for 3D Object Detection [138.2892824662943]
A promising solution is to make better use of the synthetic dataset, which consists of CAD object models, to boost the learning on real datasets.
Recent work on 3D pre-training exhibits failure when transfer features learned on synthetic objects to other real-world applications.
In this work, we put forward a new method called RandomRooms to accomplish this objective.
arXiv Detail & Related papers (2021-08-17T17:56:12Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z) - Learning Object Placements For Relational Instructions by Hallucinating
Scene Representations [26.897316325189205]
We present a convolutional neural network for estimating pixelwise object placement probabilities for a set of spatial relations from a single input image.
Our method does not require ground truth data for the pixelwise relational probabilities or 3D models of the objects.
Results obtained using real-world data and human-robot experiments demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2020-01-23T12:58:50Z) - Learning to Move with Affordance Maps [57.198806691838364]
The ability to autonomously explore and navigate a physical space is a fundamental requirement for virtually any mobile autonomous agent.
Traditional SLAM-based approaches for exploration and navigation largely focus on leveraging scene geometry.
We show that learned affordance maps can be used to augment traditional approaches for both exploration and navigation, providing significant improvements in performance.
arXiv Detail & Related papers (2020-01-08T04:05:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.