Representation, learning, and planning algorithms for geometric task and
motion planning
- URL: http://arxiv.org/abs/2203.04605v1
- Date: Wed, 9 Mar 2022 09:47:01 GMT
- Title: Representation, learning, and planning algorithms for geometric task and
motion planning
- Authors: Beomjoon Kim, Luke Shimanuki, Leslie Pack Kaelbling, Tom\'as
Lozano-P\'erez
- Abstract summary: We present a framework for learning to guide geometric task and motion planning (GTAMP)
GTAMP is a subclass of task and motion planning in which the goal is to move multiple objects to target regions among movable obstacles.
A standard graph search algorithm is not directly applicable, because GTAMP problems involve hybrid search spaces and expensive action feasibility checks.
- Score: 24.862289058632186
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present a framework for learning to guide geometric task and motion
planning (GTAMP). GTAMP is a subclass of task and motion planning in which the
goal is to move multiple objects to target regions among movable obstacles. A
standard graph search algorithm is not directly applicable, because GTAMP
problems involve hybrid search spaces and expensive action feasibility checks.
To handle this, we introduce a novel planner that extends basic heuristic
search with random sampling and a heuristic function that prioritizes
feasibility checking on promising state action pairs. The main drawback of such
pure planners is that they lack the ability to learn from planning experience
to improve their efficiency. We propose two learning algorithms to address
this. The first is an algorithm for learning a rank function that guides the
discrete task level search, and the second is an algorithm for learning a
sampler that guides the continuous motionlevel search. We propose design
principles for designing data efficient algorithms for learning from planning
experience and representations for effective generalization. We evaluate our
framework in challenging GTAMP problems, and show that we can improve both
planning and data efficiency
Related papers
- Reinforcement Learning with Success Induced Task Prioritization [68.8204255655161]
We introduce Success Induced Task Prioritization (SITP), a framework for automatic curriculum learning.
The algorithm selects the order of tasks that provide the fastest learning for agents.
We demonstrate that SITP matches or surpasses the results of other curriculum design methods.
arXiv Detail & Related papers (2022-12-30T12:32:43Z) - PALMER: Perception-Action Loop with Memory for Long-Horizon Planning [1.5469452301122177]
We introduce a general-purpose planning algorithm called PALMER.
Palmer combines classical sampling-based planning algorithms with learning-based perceptual representations.
This creates a tight feedback loop between representation learning, memory, reinforcement learning, and sampling-based planning.
arXiv Detail & Related papers (2022-12-08T22:11:49Z) - Sequence-Based Plan Feasibility Prediction for Efficient Task and Motion
Planning [36.300564378022315]
We present a learning-enabled Task and Motion Planning (TAMP) algorithm for solving mobile manipulation problems in environments with many articulated and movable obstacles.
The core of our algorithm is PIGINet, a novel Transformer-based learning method that takes in a task plan, the goal, and the initial state, and predicts the probability of finding motion trajectories associated with the task plan.
arXiv Detail & Related papers (2022-11-03T04:12:04Z) - Systematic Comparison of Path Planning Algorithms using PathBench [55.335463666037086]
Path planning is an essential component of mobile robotics.
Development of learning-based path planning algorithms has been experiencing rapid growth.
This paper presents PathBench, a platform for developing, visualizing, training, testing, and benchmarking of existing and future path planning algorithms.
arXiv Detail & Related papers (2022-03-07T01:52:57Z) - C-Planning: An Automatic Curriculum for Learning Goal-Reaching Tasks [133.40619754674066]
Goal-conditioned reinforcement learning can solve tasks in a wide range of domains, including navigation and manipulation.
We propose the distant goal-reaching task by using search at training time to automatically generate intermediate states.
E-step corresponds to planning an optimal sequence of waypoints using graph search, while the M-step aims to learn a goal-conditioned policy to reach those waypoints.
arXiv Detail & Related papers (2021-10-22T22:05:31Z) - PathBench: A Benchmarking Platform for Classical and Learned Path
Planning Algorithms [59.3879573040863]
Path planning is a key component in mobile robotics.
Few attempts have been made to benchmark the algorithms holistically or unify their interface.
This paper presents PathBench, a platform for developing, visualizing, training, testing, and benchmarking of existing and future path planning algorithms.
arXiv Detail & Related papers (2021-05-04T21:48:18Z) - Waypoint Planning Networks [66.72790309889432]
We propose a hybrid algorithm based on LSTMs with a local kernel - a classic algorithm such as A*, and a global kernel using a learned algorithm.
We compare WPN against A*, as well as related works including motion planning networks (MPNet) and value networks (VIN)
It is shown that WPN's search space is considerably less than A*, while being able to generate near optimal results.
arXiv Detail & Related papers (2021-05-01T18:02:01Z) - Learning Symbolic Operators for Task and Motion Planning [29.639902380586253]
integrated task and motion planners (TAMP) handle the complex interaction between motion-level decisions and task-level plan feasibility.
TAMP approaches rely on domain-specific symbolic operators to guide the task-level search, making planning efficient.
We propose a bottom-up relational learning method for operator learning and show how the learned operators can be used for planning in a TAMP system.
arXiv Detail & Related papers (2021-02-28T19:08:56Z) - Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning [78.65083326918351]
We consider alternatives to an implicit sequential planning assumption.
We propose Divide-and-Conquer Monte Carlo Tree Search (DC-MCTS) for approximating the optimal plan.
We show that this algorithmic flexibility over planning order leads to improved results in navigation tasks in grid-worlds.
arXiv Detail & Related papers (2020-04-23T18:08:58Z) - Flexible and Efficient Long-Range Planning Through Curious Exploration [13.260508939271764]
We show that the Curious Sample Planner can efficiently discover temporally-extended plans for solving a wide range of physically realistic 3D tasks.
In contrast, standard planning and learning methods often fail to solve these tasks at all or do so only with a huge and highly variable number of training samples.
arXiv Detail & Related papers (2020-04-22T21:47:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.