Safe Learning of PDDL Domains with Conditional Effects -- Extended Version
- URL: http://arxiv.org/abs/2403.15251v1
- Date: Fri, 22 Mar 2024 14:49:49 GMT
- Title: Safe Learning of PDDL Domains with Conditional Effects -- Extended Version
- Authors: Argaman Mordoch, Enrico Scala, Roni Stern, Brendan Juba,
- Abstract summary: We show that Conditional-SAM can be used to solve perfectly most of the test set problems in most of the experimented domains.
Our results show that the action models learned by Conditional-SAM can be used to solve perfectly most of the test set problems.
- Score: 27.05167679870857
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Powerful domain-independent planners have been developed to solve various types of planning problems. These planners often require a model of the acting agent's actions, given in some planning domain description language. Manually designing such an action model is a notoriously challenging task. An alternative is to automatically learn action models from observation. Such an action model is called safe if every plan created with it is consistent with the real, unknown action model. Algorithms for learning such safe action models exist, yet they cannot handle domains with conditional or universal effects, which are common constructs in many planning problems. We prove that learning non-trivial safe action models with conditional effects may require an exponential number of samples. Then, we identify reasonable assumptions under which such learning is tractable and propose SAM Learning of Conditional Effects (Conditional-SAM), the first algorithm capable of doing so. We analyze Conditional-SAM theoretically and evaluate it experimentally. Our results show that the action models learned by Conditional-SAM can be used to solve perfectly most of the test set problems in most of the experimented domains.
Related papers
- Action Model Learning with Guarantees [5.524804393257921]
We develop a theory for action model learning based on version spaces that interprets the task as search for hypothesis that are consistent with the learning examples.
Our theoretical findings are instantiated in an online algorithm that maintains a compact representation of all solutions of the problem.
arXiv Detail & Related papers (2024-04-15T10:01:43Z) - Enhancing Numeric-SAM for Learning with Few Observations [13.41686187754024]
We propose an enhanced version of Safe Action Models Learning (N-SAM) that always returns an action model where every observed action is applicable at least in some state.
N-SAM* does so without compromising the safety of the returned action model.
An empirical study on a set of benchmark domains shows that the action models returned by N-SAM* enable solving significantly more problems than the action models returned by N-SAM.
arXiv Detail & Related papers (2023-12-17T12:50:10Z) - DiMSam: Diffusion Models as Samplers for Task and Motion Planning under Partial Observability [58.75803543245372]
Task and Motion Planning (TAMP) approaches are suited for planning multi-step autonomous robot manipulation.
We propose to overcome these limitations by composing diffusion models using a TAMP system.
We show how the combination of classical TAMP, generative modeling, and latent embedding enables multi-step constraint-based reasoning.
arXiv Detail & Related papers (2023-06-22T20:40:24Z) - SimSCOOD: Systematic Analysis of Out-of-Distribution Generalization in
Fine-tuned Source Code Models [58.78043959556283]
We study the behaviors of models under different fine-tuning methodologies, including full fine-tuning and Low-Rank Adaptation (LoRA) fine-tuning methods.
Our analysis uncovers that LoRA fine-tuning consistently exhibits significantly better OOD generalization performance than full fine-tuning across various scenarios.
arXiv Detail & Related papers (2022-10-10T16:07:24Z) - Evaluating model-based planning and planner amortization for continuous
control [79.49319308600228]
We take a hybrid approach, combining model predictive control (MPC) with a learned model and model-free policy learning.
We find that well-tuned model-free agents are strong baselines even for high DoF control problems.
We show that it is possible to distil a model-based planner into a policy that amortizes the planning without any loss of performance.
arXiv Detail & Related papers (2021-10-07T12:00:40Z) - Safe Learning of Lifted Action Models [46.65973550325976]
We propose an algorithm for solving the model-free planning problem in classical planning.
The number of trajectories needed to solve future problems with high probability is linear in the potential size of the domain model.
arXiv Detail & Related papers (2021-07-09T01:24:01Z) - Deliberative Acting, Online Planning and Learning with Hierarchical
Operational Models [5.597986898418404]
In AI research, a plan of action has typically used descriptive models of the actions that abstractly specify what might happen as a result of an action.
executing the planned actions has needed operational models, in which rich computational control structures and closed-loop online decision-making are used.
We implement an integrated acting and planning system in which both planning and acting use the same operational models.
arXiv Detail & Related papers (2020-10-02T14:50:05Z) - Goal-Aware Prediction: Learning to Model What Matters [105.43098326577434]
One of the fundamental challenges in using a learned forward dynamics model is the mismatch between the objective of the learned model and that of the downstream planner or policy.
We propose to direct prediction towards task relevant information, enabling the model to be aware of the current task and encouraging it to only model relevant quantities of the state space.
We find that our method more effectively models the relevant parts of the scene conditioned on the goal, and as a result outperforms standard task-agnostic dynamics models and model-free reinforcement learning.
arXiv Detail & Related papers (2020-07-14T16:42:59Z) - STRIPS Action Discovery [67.73368413278631]
Recent approaches have shown the success of classical planning at synthesizing action models even when all intermediate states are missing.
We propose a new algorithm to unsupervisedly synthesize STRIPS action models with a classical planner when action signatures are unknown.
arXiv Detail & Related papers (2020-01-30T17:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.