Effective Baselines for Multiple Object Rearrangement Planning in
Partially Observable Mapped Environments
- URL: http://arxiv.org/abs/2301.09854v1
- Date: Tue, 24 Jan 2023 08:03:34 GMT
- Title: Effective Baselines for Multiple Object Rearrangement Planning in
Partially Observable Mapped Environments
- Authors: Engin Tekin, Elaheh Barati, Nitin Kamra, Ruta Desai
- Abstract summary: This paper aims to enable home-assistive intelligent agents to efficiently plan for rearrangement under partial observability.
We investigate monolithic and modular deep reinforcement learning (DRL) methods for planning in our setting.
We find that monolithic DRL methods do not succeed at long-horizon planning needed for multi-object rearrangement.
We also show that our greedy modular agents are empirically optimal when the objects that need to be rearranged are uniformly distributed in the environment.
- Score: 5.32429768581469
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Many real-world tasks, from house-cleaning to cooking, can be formulated as
multi-object rearrangement problems -- where an agent needs to get specific
objects into appropriate goal states. For such problems, we focus on the
setting that assumes a pre-specified goal state, availability of perfect
manipulation and object recognition capabilities, and a static map of the
environment but unknown initial location of objects to be rearranged. Our goal
is to enable home-assistive intelligent agents to efficiently plan for
rearrangement under such partial observability. This requires efficient
trade-offs between exploration of the environment and planning for
rearrangement, which is challenging because of long-horizon nature of the
problem. To make progress on this problem, we first analyze the effects of
various factors such as number of objects and receptacles, agent carrying
capacity, environment layouts etc. on exploration and planning for
rearrangement using classical methods. We then investigate both monolithic and
modular deep reinforcement learning (DRL) methods for planning in our setting.
We find that monolithic DRL methods do not succeed at long-horizon planning
needed for multi-object rearrangement. Instead, modular greedy approaches
surprisingly perform reasonably well and emerge as competitive baselines for
planning with partial observability in multi-object rearrangement problems. We
also show that our greedy modular agents are empirically optimal when the
objects that need to be rearranged are uniformly distributed in the environment
-- thereby contributing baselines with strong performance for future work on
multi-object rearrangement planning in partially observable settings.
Related papers
- ET-Plan-Bench: Embodied Task-level Planning Benchmark Towards Spatial-Temporal Cognition with Foundation Models [39.606908488885125]
ET-Plan-Bench is a benchmark for embodied task planning using Large Language Models (LLMs)
It features a controllable and diverse set of embodied tasks varying in different levels of difficulties and complexities.
Our benchmark distinguishes itself as a large-scale, quantifiable, highly automated, and fine-grained diagnostic framework.
arXiv Detail & Related papers (2024-10-02T19:56:38Z) - Unified Task and Motion Planning using Object-centric Abstractions of
Motion Constraints [56.283944756315066]
We propose an alternative TAMP approach that unifies task and motion planning into a single search.
Our approach is based on an object-centric abstraction of motion constraints that permits leveraging the computational efficiency of off-the-shelf AI search to yield physically feasible plans.
arXiv Detail & Related papers (2023-12-29T14:00:20Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered
Environments [8.15681999722805]
This paper proposes a learning-based framework for multi-agent object rearrangement planning.
It addresses the challenges of task sequencing and path planning in complex environments.
arXiv Detail & Related papers (2023-06-10T23:53:28Z) - AdaPlanner: Adaptive Planning from Feedback with Language Models [56.367020818139665]
Large language models (LLMs) have recently demonstrated the potential in acting as autonomous agents for sequential decision-making tasks.
We propose a closed-loop approach, AdaPlanner, which allows the LLM agent to refine its self-generated plan adaptively in response to environmental feedback.
To mitigate hallucination, we develop a code-style LLM prompt structure that facilitates plan generation across a variety of tasks, environments, and agent capabilities.
arXiv Detail & Related papers (2023-05-26T05:52:27Z) - Online Tool Selection with Learned Grasp Prediction Models [15.915410623440875]
Deep learning-based grasp prediction models have become an industry standard for robotic bin-picking systems.
To maximize pick success, production environments are often equipped with several end-effector tools that can be swapped on-the-fly, based on the target object.
We show how to choose the order of grasps to perform, and corresponding tool-change actions, to improve system throughput.
arXiv Detail & Related papers (2023-02-15T20:52:38Z) - Task Allocation for Multi-Robot Task and Motion Planning: a case for
Object Picking in Cluttered Workspaces [1.3535770763481902]
We present an integrated multi-robot task and motion planning approach.
It is capable of handling tasks which involve an unknown number of object re-arrangements.
We demonstrate our results with experiments in simulation on two Franka Emika manipulators.
arXiv Detail & Related papers (2021-10-08T12:36:43Z) - Hierarchical Policy for Non-prehensile Multi-object Rearrangement with
Deep Reinforcement Learning and Monte Carlo Tree Search [30.31462739429364]
We propose a hierarchical policy to divide and conquer for non-prehensile multi-object rearrangement.
In the high-level policy, the Monte Carlo Tree Search efficiently searches for the optimal rearrangement sequence among multiple objects.
In the low-level policy, the robot plans the paths according to the order of path primitives and manipulates the objects to approach the goal poses one by one.
arXiv Detail & Related papers (2021-09-18T17:24:37Z) - NeRP: Neural Rearrangement Planning for Unknown Objects [49.191284597526]
We propose NeRP (Neural Rearrangement Planning), a deep learning based approach for multi-step neural object rearrangement planning.
NeRP works with never-before-seen objects, that is trained on simulation data, and generalizes to the real world.
arXiv Detail & Related papers (2021-06-02T17:56:27Z) - PackIt: A Virtual Environment for Geometric Planning [68.79816936618454]
PackIt is a virtual environment to evaluate and potentially learn the ability to do geometric planning.
We construct a set of challenging packing tasks using an evolutionary algorithm.
arXiv Detail & Related papers (2020-07-21T22:51:17Z) - Bottom-up mechanism and improved contract net protocol for the dynamic
task planning of heterogeneous Earth observation resources [61.75759893720484]
Earth observation resources are becoming increasingly indispensable in disaster relief, damage assessment and related domains.
Many unpredicted factors, such as the change of observation task requirements, to the occurring of bad weather and resource failures, may cause the scheduled observation scheme to become infeasible.
A bottom-up distributed coordinated framework together with an improved contract net are proposed to facilitate the dynamic task replanning for heterogeneous Earth observation resources.
arXiv Detail & Related papers (2020-07-13T03:51:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.