Task Scoping: Generating Task-Specific Abstractions for Planning
- URL: http://arxiv.org/abs/2010.08869v2
- Date: Tue, 11 May 2021 02:44:38 GMT
- Title: Task Scoping: Generating Task-Specific Abstractions for Planning
- Authors: Nishanth Kumar, Michael Fishman, Natasha Danas, Michael Littman,
Stefanie Tellex, George Konidaris
- Abstract summary: Planning to solve any specific task using an open-scope world model is computationally intractable.
We propose task scoping: a method that exploits knowledge of the initial condition, goal condition, and transition-dynamics structure of a task.
We prove that task scoping never deletes relevant factors or actions, characterize its computational complexity, and characterize the planning problems for which it is especially useful.
- Score: 19.411900372400183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A generally intelligent agent requires an open-scope world model: one rich
enough to tackle any of the wide range of tasks it may be asked to solve over
its operational lifetime. Unfortunately, planning to solve any specific task
using such a rich model is computationally intractable - even for
state-of-the-art methods - due to the many states and actions that are
necessarily present in the model but irrelevant to that problem. We propose
task scoping: a method that exploits knowledge of the initial condition, goal
condition, and transition-dynamics structure of a task to automatically and
efficiently prune provably irrelevant factors and actions from a planning
problem, which can dramatically decrease planning time. We prove that task
scoping never deletes relevant factors or actions, characterize its
computational complexity, and characterize the planning problems for which it
is especially useful. Finally, we empirically evaluate task scoping on a
variety of domains and demonstrate that using it as a pre-planning step can
reduce the state-action space of various planning problems by orders of
magnitude and speed up planning. When applied to a complex Minecraft domain,
our approach speeds up a state-of-the-art planner by 30 times, including the
time required for task scoping itself.
Related papers
- Propose, Assess, Search: Harnessing LLMs for Goal-Oriented Planning in Instructional Videos [48.15438373870542]
VidAssist is an integrated framework designed for zero/few-shot goal-oriented planning in instructional videos.
It employs a breadth-first search algorithm for optimal plan generation.
Experiments demonstrate that VidAssist offers a unified framework for different goal-oriented planning setups.
arXiv Detail & Related papers (2024-09-30T17:57:28Z) - Unified Task and Motion Planning using Object-centric Abstractions of
Motion Constraints [56.283944756315066]
We propose an alternative TAMP approach that unifies task and motion planning into a single search.
Our approach is based on an object-centric abstraction of motion constraints that permits leveraging the computational efficiency of off-the-shelf AI search to yield physically feasible plans.
arXiv Detail & Related papers (2023-12-29T14:00:20Z) - Learning adaptive planning representations with natural language
guidance [90.24449752926866]
This paper describes Ada, a framework for automatically constructing task-specific planning representations.
Ada interactively learns a library of planner-compatible high-level action abstractions and low-level controllers adapted to a particular domain of planning tasks.
arXiv Detail & Related papers (2023-12-13T23:35:31Z) - Planning as In-Painting: A Diffusion-Based Embodied Task Planning
Framework for Environments under Uncertainty [56.30846158280031]
Task planning for embodied AI has been one of the most challenging problems.
We propose a task-agnostic method named 'planning as in-painting'
The proposed framework achieves promising performances in various embodied AI tasks.
arXiv Detail & Related papers (2023-12-02T10:07:17Z) - Robot Task Planning Based on Large Language Model Representing Knowledge
with Directed Graph Structures [2.3698227130544547]
We propose a task planning method that combines human expertise with an LLM and have designed an LLM prompt template, Think_Net_Prompt.
We further propose a method to progressively decompose tasks and generate a task tree to reduce the planning volume for each task.
arXiv Detail & Related papers (2023-06-08T13:10:00Z) - Optimal task and motion planning and execution for human-robot
multi-agent systems in dynamic environments [54.39292848359306]
We propose a combined task and motion planning approach to optimize sequencing, assignment, and execution of tasks.
The framework relies on decoupling tasks and actions, where an action is one possible geometric realization of a symbolic task.
We demonstrate the approach effectiveness in a collaborative manufacturing scenario, in which a robotic arm and a human worker shall assemble a mosaic.
arXiv Detail & Related papers (2023-03-27T01:50:45Z) - POMRL: No-Regret Learning-to-Plan with Increasing Horizons [43.693739167594295]
We study the problem of planning under model uncertainty in an online meta-reinforcement learning setting.
We propose an algorithm to meta-learn the underlying structure across tasks, utilize it to plan in each task, and upper-bound the regret of the planning loss.
arXiv Detail & Related papers (2022-12-30T03:09:45Z) - Anytime Stochastic Task and Motion Policies [12.72186877599064]
We present a new approach for integrated task and motion planning in settings.
Our algorithm is probabilistically complete and can compute feasible solution policies in an anytime fashion.
arXiv Detail & Related papers (2021-08-28T00:23:39Z) - Planning with Learned Object Importance in Large Problem Instances using
Graph Neural Networks [28.488201307961624]
Real-world planning problems often involve hundreds or even thousands of objects.
We propose a graph neural network architecture for predicting object importance in a single inference pass.
Our approach treats the planner and transition model as black boxes, and can be used with any off-the-shelf planner.
arXiv Detail & Related papers (2020-09-11T18:55:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.