Simultaneous Navigation and Construction Benchmarking Environments
- URL: http://arxiv.org/abs/2103.16732v1
- Date: Wed, 31 Mar 2021 00:05:54 GMT
- Title: Simultaneous Navigation and Construction Benchmarking Environments
- Authors: Wenyu Han, Chen Feng, Haoran Wu, Alexander Gao, Armand Jordana, Dong
Liu, Lerrel Pinto, Ludovic Righetti
- Abstract summary: We need intelligent robots for mobile construction, the process of navigating in an environment and modifying its structure according to a geometric design.
In this task, a major robot vision and learning challenge is how to exactly achieve the design without GPS.
We benchmark the performance of a handcrafted policy with basic localization and planning, and state-of-the-art deep reinforcement learning methods.
- Score: 73.0706832393065
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We need intelligent robots for mobile construction, the process of navigating
in an environment and modifying its structure according to a geometric design.
In this task, a major robot vision and learning challenge is how to exactly
achieve the design without GPS, due to the difficulty caused by the
bi-directional coupling of accurate robot localization and navigation together
with strategic environment manipulation. However, many existing robot vision
and learning tasks such as visual navigation and robot manipulation address
only one of these two coupled aspects. To stimulate the pursuit of a generic
and adaptive solution, we reasonably simplify mobile construction as a
partially observable Markov decision process (POMDP) in 1/2/3D grid worlds and
benchmark the performance of a handcrafted policy with basic localization and
planning, and state-of-the-art deep reinforcement learning (RL) methods. Our
extensive experiments show that the coupling makes this problem very
challenging for those methods, and emphasize the need for novel task-specific
solutions.
Related papers
- Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Evolutionary Swarm Robotics: Dynamic Subgoal-Based Path Formation and
Task Allocation for Exploration and Navigation in Unknown Environments [0.0]
The paper presents a method called the sub-goal-based path formation, which establishes a path between two different locations by exploiting visually connected sub-goals.
The paper tackles the problem of inter-collision (traffic) among a large number of robots engaged in path formation, which negatively impacts the performance of the sub-goal-based method.
A task allocation strategy is proposed, leveraging local communication protocols and light signal-based communication.
arXiv Detail & Related papers (2023-12-27T15:13:56Z) - Learning Hierarchical Interactive Multi-Object Search for Mobile
Manipulation [10.21450780640562]
We introduce a novel interactive multi-object search task in which a robot has to open doors to navigate rooms and search inside cabinets and drawers to find target objects.
These new challenges require combining manipulation and navigation skills in unexplored environments.
We present HIMOS, a hierarchical reinforcement learning approach that learns to compose exploration, navigation, and manipulation skills.
arXiv Detail & Related papers (2023-07-12T12:25:33Z) - Robot path planning using deep reinforcement learning [0.0]
Reinforcement learning methods offer an alternative to map-free navigation tasks.
Deep reinforcement learning agents are implemented for both the obstacle avoidance and the goal-oriented navigation task.
An analysis of the changes in the behaviour and performance of the agents caused by modifications in the reward function is conducted.
arXiv Detail & Related papers (2023-02-17T20:08:59Z) - ViNG: Learning Open-World Navigation with Visual Goals [82.84193221280216]
We propose a learning-based navigation system for reaching visually indicated goals.
We show that our system, which we call ViNG, outperforms previously-proposed methods for goal-conditioned reinforcement learning.
We demonstrate ViNG on a number of real-world applications, such as last-mile delivery and warehouse inspection.
arXiv Detail & Related papers (2020-12-17T18:22:32Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Mobile Robot Path Planning in Dynamic Environments through Globally
Guided Reinforcement Learning [12.813442161633116]
We introduce a globally guided learning reinforcement approach (G2RL) to solve the multi-robot planning problem.
G2RL incorporates a novel path reward structure that generalizes to arbitrary environments.
We evaluate our method across different map types, obstacle densities and the number of robots.
arXiv Detail & Related papers (2020-05-11T20:42:29Z) - SAPIEN: A SimulAted Part-based Interactive ENvironment [77.4739790629284]
SAPIEN is a realistic and physics-rich simulated environment that hosts a large-scale set for articulated objects.
We evaluate state-of-the-art vision algorithms for part detection and motion attribute recognition as well as demonstrate robotic interaction tasks.
arXiv Detail & Related papers (2020-03-19T00:11:34Z) - BADGR: An Autonomous Self-Supervised Learning-Based Navigation System [158.6392333480079]
BadGR is an end-to-end learning-based mobile robot navigation system.
It can be trained with self-supervised off-policy data gathered in real-world environments.
BadGR can navigate in real-world urban and off-road environments with geometrically distracting obstacles.
arXiv Detail & Related papers (2020-02-13T18:40:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.