Learning Symbolic Operators for Task and Motion Planning
- URL: http://arxiv.org/abs/2103.00589v1
- Date: Sun, 28 Feb 2021 19:08:56 GMT
- Title: Learning Symbolic Operators for Task and Motion Planning
- Authors: Tom Silver, Rohan Chitnis, Joshua Tenenbaum, Leslie Pack Kaelbling,
Tomas Lozano-Perez
- Abstract summary: integrated task and motion planners (TAMP) handle the complex interaction between motion-level decisions and task-level plan feasibility.
TAMP approaches rely on domain-specific symbolic operators to guide the task-level search, making planning efficient.
We propose a bottom-up relational learning method for operator learning and show how the learned operators can be used for planning in a TAMP system.
- Score: 29.639902380586253
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Robotic planning problems in hybrid state and action spaces can be solved by
integrated task and motion planners (TAMP) that handle the complex interaction
between motion-level decisions and task-level plan feasibility. TAMP approaches
rely on domain-specific symbolic operators to guide the task-level search,
making planning efficient. In this work, we formalize and study the problem of
operator learning for TAMP. Central to this study is the view that operators
define a lossy abstraction of the transition model of the underlying domain. We
then propose a bottom-up relational learning method for operator learning and
show how the learned operators can be used for planning in a TAMP system.
Experimentally, we provide results in three domains, including long-horizon
robotic planning tasks. We find our approach to substantially outperform
several baselines, including three graph neural network-based model-free
approaches based on recent work. Video: https://youtu.be/iVfpX9BpBRo
Related papers
- A Survey of Optimization-based Task and Motion Planning: From Classical To Learning Approaches [15.136760934936381]
Task and Motion Planning (TAMP) integrates high-level task planning and low-level motion planning to equip robots with the autonomy to reason over long-horizon, dynamic tasks.
This survey provides a comprehensive review on optimization-based TAMP, covering (i) planning domain representations, (ii) individual solution strategies for components, including AI planning and trajectory optimization (TO), and (iii) the dynamic interplay between logic-based task planning and model-based TO.
arXiv Detail & Related papers (2024-04-03T15:38:36Z) - LLM3:Large Language Model-based Task and Motion Planning with Motion Failure Reasoning [78.2390460278551]
Conventional Task and Motion Planning (TAMP) approaches rely on manually crafted interfaces connecting symbolic task planning with continuous motion generation.
Here, we present LLM3, a novel Large Language Model (LLM)-based TAMP framework featuring a domain-independent interface.
Specifically, we leverage the powerful reasoning and planning capabilities of pre-trained LLMs to propose symbolic action sequences and select continuous action parameters for motion planning.
arXiv Detail & Related papers (2024-03-18T08:03:47Z) - Unified Task and Motion Planning using Object-centric Abstractions of
Motion Constraints [56.283944756315066]
We propose an alternative TAMP approach that unifies task and motion planning into a single search.
Our approach is based on an object-centric abstraction of motion constraints that permits leveraging the computational efficiency of off-the-shelf AI search to yield physically feasible plans.
arXiv Detail & Related papers (2023-12-29T14:00:20Z) - DiMSam: Diffusion Models as Samplers for Task and Motion Planning under
Partial Observability [50.38132214102161]
Task and Motion Planning (TAMP) approaches are effective at planning long-horizon autonomous robot manipulation.
We propose to overcome these limitations by leveraging deep generative modeling.
We show how the combination of classical TAMP, generative learning, and latent embeddings enables long-horizon constraint-based reasoning.
arXiv Detail & Related papers (2023-06-22T20:40:24Z) - Learning to Reason over Scene Graphs: A Case Study of Finetuning GPT-2
into a Robot Language Model for Grounded Task Planning [45.51792981370957]
We investigate the applicability of a smaller class of large language models (LLMs) in robotic task planning by learning to decompose tasks into subgoal specifications for a planner to execute sequentially.
Our method grounds the input of the LLM on the domain that is represented as a scene graph, enabling it to translate human requests into executable robot plans.
Our findings suggest that the knowledge stored in an LLM can be effectively grounded to perform long-horizon task planning, demonstrating the promising potential for the future application of neuro-symbolic planning methods in robotics.
arXiv Detail & Related papers (2023-05-12T18:14:32Z) - Hierarchical Imitation Learning with Vector Quantized Models [77.67190661002691]
We propose to use reinforcement learning to identify subgoals in expert trajectories.
We build a vector-quantized generative model for the identified subgoals to perform subgoal-level planning.
In experiments, the algorithm excels at solving complex, long-horizon decision-making problems outperforming state-of-the-art.
arXiv Detail & Related papers (2023-01-30T15:04:39Z) - Sequence-Based Plan Feasibility Prediction for Efficient Task and Motion
Planning [36.300564378022315]
We present a learning-enabled Task and Motion Planning (TAMP) algorithm for solving mobile manipulation problems in environments with many articulated and movable obstacles.
The core of our algorithm is PIGINet, a novel Transformer-based learning method that takes in a task plan, the goal, and the initial state, and predicts the probability of finding motion trajectories associated with the task plan.
arXiv Detail & Related papers (2022-11-03T04:12:04Z) - Learning Efficient Abstract Planning Models that Choose What to Predict [28.013014215441505]
We show that existing symbolic operator learning approaches fall short in many robotics domains.
This is primarily because they attempt to learn operators that exactly predict all observed changes in the abstract state.
We propose to learn operators that 'choose what to predict' by only modelling changes necessary for abstract planning to achieve specified goals.
arXiv Detail & Related papers (2022-08-16T13:12:59Z) - Learning Neuro-Symbolic Skills for Bilevel Planning [63.388694268198655]
Decision-making is challenging in robotics environments with continuous object-centric states, continuous actions, long horizons, and sparse feedback.
Hierarchical approaches, such as task and motion planning (TAMP), address these challenges by decomposing decision-making into two or more levels of abstraction.
Our main contribution is a method for learning parameterized polices in combination with operators and samplers.
arXiv Detail & Related papers (2022-06-21T19:01:19Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.