Active Learning of Abstract Plan Feasibility
- URL: http://arxiv.org/abs/2107.00683v1
- Date: Thu, 1 Jul 2021 18:17:01 GMT
- Title: Active Learning of Abstract Plan Feasibility
- Authors: Michael Noseworthy, Caris Moses, Isaiah Brand, Sebastian Castro,
Leslie Kaelbling, Tom\'as Lozano-P\'erez, Nicholas Roy
- Abstract summary: We present an active learning approach to efficiently acquire an APF predictor through task-independent, curious exploration on a robot.
We leverage an infeasible subsequence property to prune candidate plans in the active learning strategy, allowing our system to learn from less data.
In a stacking domain where objects have non-uniform mass distributions, we show that our system permits real robot learning of an APF model in four hundred self-supervised interactions.
- Score: 17.689758291966502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Long horizon sequential manipulation tasks are effectively addressed
hierarchically: at a high level of abstraction the planner searches over
abstract action sequences, and when a plan is found, lower level motion plans
are generated. Such a strategy hinges on the ability to reliably predict that a
feasible low level plan will be found which satisfies the abstract plan.
However, computing Abstract Plan Feasibility (APF) is difficult because the
outcome of a plan depends on real-world phenomena that are difficult to model,
such as noise in estimation and execution. In this work, we present an active
learning approach to efficiently acquire an APF predictor through
task-independent, curious exploration on a robot. The robot identifies plans
whose outcomes would be informative about APF, executes those plans, and learns
from their successes or failures. Critically, we leverage an infeasible
subsequence property to prune candidate plans in the active learning strategy,
allowing our system to learn from less data. We evaluate our strategy in
simulation and on a real Franka Emika Panda robot with integrated perception,
experimentation, planning, and execution. In a stacking domain where objects
have non-uniform mass distributions, we show that our system permits real robot
learning of an APF model in four hundred self-supervised interactions, and that
our learned model can be used effectively in multiple downstream tasks.
Related papers
- KnowAgent: Knowledge-Augmented Planning for LLM-Based Agents [54.09074527006576]
Large Language Models (LLMs) have demonstrated great potential in complex reasoning tasks, yet they fall short when tackling more sophisticated challenges.
This inadequacy primarily stems from the lack of built-in action knowledge in language agents.
We introduce KnowAgent, a novel approach designed to enhance the planning capabilities of LLMs by incorporating explicit action knowledge.
arXiv Detail & Related papers (2024-03-05T16:39:12Z) - Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Learning to Reason over Scene Graphs: A Case Study of Finetuning GPT-2
into a Robot Language Model for Grounded Task Planning [45.51792981370957]
We investigate the applicability of a smaller class of large language models (LLMs) in robotic task planning by learning to decompose tasks into subgoal specifications for a planner to execute sequentially.
Our method grounds the input of the LLM on the domain that is represented as a scene graph, enabling it to translate human requests into executable robot plans.
Our findings suggest that the knowledge stored in an LLM can be effectively grounded to perform long-horizon task planning, demonstrating the promising potential for the future application of neuro-symbolic planning methods in robotics.
arXiv Detail & Related papers (2023-05-12T18:14:32Z) - Efficient Learning of High Level Plans from Play [57.29562823883257]
We present Efficient Learning of High-Level Plans from Play (ELF-P), a framework for robotic learning that bridges motion planning and deep RL.
We demonstrate that ELF-P has significantly better sample efficiency than relevant baselines over multiple realistic manipulation tasks.
arXiv Detail & Related papers (2023-03-16T20:09:47Z) - Hierarchical Imitation Learning with Vector Quantized Models [77.67190661002691]
We propose to use reinforcement learning to identify subgoals in expert trajectories.
We build a vector-quantized generative model for the identified subgoals to perform subgoal-level planning.
In experiments, the algorithm excels at solving complex, long-horizon decision-making problems outperforming state-of-the-art.
arXiv Detail & Related papers (2023-01-30T15:04:39Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - Inventing Relational State and Action Abstractions for Effective and
Efficient Bilevel Planning [26.715198108255162]
We develop a novel framework for learning state and action abstractions.
We learn relational, neuro-symbolic abstractions that generalize over object identities and numbers.
We show that our learned abstractions are able to quickly solve held-out tasks of longer horizons.
arXiv Detail & Related papers (2022-03-17T22:13:09Z) - Using Deep Learning to Bootstrap Abstractions for Hierarchical Robot
Planning [27.384742641275228]
We present a new approach for bootstrapping the entire hierarchical planning process.
It shows how abstract states and actions for new environments can be computed automatically.
It uses the learned abstractions in a novel multi-source bi-directional hierarchical robot planning algorithm.
arXiv Detail & Related papers (2022-02-02T08:11:20Z) - Self-Imitation Learning by Planning [3.996275177789895]
Imitation learning (IL) enables robots to acquire skills quickly by transferring expert knowledge.
In long-horizon motion planning tasks, a challenging problem in deploying IL and RL methods is how to generate and collect massive, broadly distributed data.
We propose self-imitation learning by planning (SILP), where demonstration data are collected automatically by planning on the visited states from the current policy.
SILP is inspired by the observation that successfully visited states in the early reinforcement learning stage are collision-free nodes in the graph-search based motion planner.
arXiv Detail & Related papers (2021-03-25T13:28:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.