Learning Type-Generalized Actions for Symbolic Planning
- URL: http://arxiv.org/abs/2308.04867v1
- Date: Wed, 9 Aug 2023 11:01:46 GMT
- Title: Learning Type-Generalized Actions for Symbolic Planning
- Authors: Daniel Tanneberg, Michael Gienger
- Abstract summary: We propose a novel concept to generalize symbolic actions using a given entity hierarchy.
In a simulated grid-based kitchen environment, we show that type-generalized actions can be learned from few observations and generalize to novel situations.
- Score: 4.670305538969915
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Symbolic planning is a powerful technique to solve complex tasks that require
long sequences of actions and can equip an intelligent agent with complex
behavior. The downside of this approach is the necessity for suitable symbolic
representations describing the state of the environment as well as the actions
that can change it. Traditionally such representations are carefully
hand-designed by experts for distinct problem domains, which limits their
transferability to different problems and environment complexities. In this
paper, we propose a novel concept to generalize symbolic actions using a given
entity hierarchy and observed similar behavior. In a simulated grid-based
kitchen environment, we show that type-generalized actions can be learned from
few observations and generalize to novel situations. Incorporating an
additional on-the-fly generalization mechanism during planning, unseen task
combinations, involving longer sequences, novel entities and unexpected
environment behavior, can be solved.
Related papers
- AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Optimal task and motion planning and execution for human-robot
multi-agent systems in dynamic environments [54.39292848359306]
We propose a combined task and motion planning approach to optimize sequencing, assignment, and execution of tasks.
The framework relies on decoupling tasks and actions, where an action is one possible geometric realization of a symbolic task.
We demonstrate the approach effectiveness in a collaborative manufacturing scenario, in which a robotic arm and a human worker shall assemble a mosaic.
arXiv Detail & Related papers (2023-03-27T01:50:45Z) - Inferring Versatile Behavior from Demonstrations by Matching Geometric
Descriptors [72.62423312645953]
Humans intuitively solve tasks in versatile ways, varying their behavior in terms of trajectory-based planning and for individual steps.
Current Imitation Learning algorithms often only consider unimodal expert demonstrations and act in a state-action-based setting.
Instead, we combine a mixture of movement primitives with a distribution matching objective to learn versatile behaviors that match the expert's behavior and versatility.
arXiv Detail & Related papers (2022-10-17T16:42:59Z) - Provably Efficient Causal Model-Based Reinforcement Learning for
Systematic Generalization [30.456180468318305]
In the sequential decision making setting, an agent aims to achieve systematic generalization over a large, possibly infinite, set of environments.
In this paper, we provide a tractable formulation of systematic generalization by employing a causal viewpoint.
Under specific structural assumptions, we provide a simple learning algorithm that guarantees any desired planning error up to an unavoidable sub-optimality term.
arXiv Detail & Related papers (2022-02-14T08:34:51Z) - Anytime Stochastic Task and Motion Policies [12.72186877599064]
We present a new approach for integrated task and motion planning in settings.
Our algorithm is probabilistically complete and can compute feasible solution policies in an anytime fashion.
arXiv Detail & Related papers (2021-08-28T00:23:39Z) - Task and Situation Structures for Service Agent Planning [13.316408384365308]
We introduce a generic structure for representing tasks, and another structure for representing situations.
Based on the two newly introduced structures, we present a methodology of situation handling that avoids hard-coding domain rules.
arXiv Detail & Related papers (2021-07-27T14:33:49Z) - CausalCity: Complex Simulations with Agency for Causal Discovery and
Reasoning [68.74447489372037]
We present a high-fidelity simulation environment that is designed for developing algorithms for causal discovery and counterfactual reasoning.
A core component of our work is to introduce textitagency, such that it is simple to define and create complex scenarios.
We perform experiments with three state-of-the-art methods to create baselines and highlight the affordances of this environment.
arXiv Detail & Related papers (2021-06-25T00:21:41Z) - Generalized Inverse Planning: Learning Lifted non-Markovian Utility for
Generalizable Task Representation [83.55414555337154]
In this work, we study learning such utility from human demonstrations.
We propose a new quest, Generalized Inverse Planning, for utility learning in this domain.
We outline a computational framework, Maximum Entropy Inverse Planning (MEIP), that learns non-Markovian utility and associated concepts in a generative manner.
arXiv Detail & Related papers (2020-11-12T21:06:26Z) - Efficient State Abstraction using Object-centered Predicates for
Manipulation Planning [86.24148040040885]
We propose an object-centered representation that permits characterizing a much wider set of possible changes in configuration spaces.
Based on this representation, we define universal planning operators for picking and placing actions that permits generating plans with geometric and force consistency.
arXiv Detail & Related papers (2020-07-16T10:52:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.