Safe-Planner: A Single-Outcome Replanner for Computing Strong Cyclic
Policies in Fully Observable Non-Deterministic Domains
- URL: http://arxiv.org/abs/2109.11471v1
- Date: Thu, 23 Sep 2021 16:20:35 GMT
- Title: Safe-Planner: A Single-Outcome Replanner for Computing Strong Cyclic
Policies in Fully Observable Non-Deterministic Domains
- Authors: Vahid Mokhtari, Ajay Suresha Sathya, Nikolaos Tsiogkas, Wilm Decre
- Abstract summary: We introduce an offline replanner, called Safe-Planner, that relies on a single-outcome determinization to compile a non-deterministic domain to a set of classical domains.
We show experimentally that this approach can allow SP to avoid generating misleading plans but to generate weak plans that directly lead to strong solutions.
- Score: 0.22940141855172028
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Replanners are efficient methods for solving non-deterministic planning
problems. Despite showing good scalability, existing replanners often fail to
solve problems involving a large number of misleading plans, i.e., weak plans
that do not lead to strong solutions, however, due to their minimal lengths,
are likely to be found at every replanning iteration. The poor performance of
replanners in such problems is due to their all-outcome determinization. That
is, when compiling from non-deterministic to classical, they include all
compiled classical operators in a single deterministic domain which leads
replanners to continually generate misleading plans. We introduce an offline
replanner, called Safe-Planner (SP), that relies on a single-outcome
determinization to compile a non-deterministic domain to a set of classical
domains, and ordering heuristics for ranking the obtained classical domains.
The proposed single-outcome determinization and the heuristics allow for
alternating between different classical domains. We show experimentally that
this approach can allow SP to avoid generating misleading plans but to generate
weak plans that directly lead to strong solutions. The experiments show that SP
outperforms state-of-the-art non-deterministic solvers by solving a broader
range of problems. We also validate the practical utility of SP in real-world
non-deterministic robotic tasks.
Related papers
- On Sequential Fault-Intolerant Process Planning [60.66853798340345]
We propose and study a planning problem we call Sequential Fault-Intolerant Process Planning (SFIPP)
SFIPP captures a reward structure common in many sequential multi-stage decision problems where the planning is deemed successful only if all stages succeed.
We design provably tight online algorithms for settings in which we need to pick between different actions with unknown success chances at each stage.
arXiv Detail & Related papers (2025-02-07T15:20:35Z) - LLM-Generated Heuristics for AI Planning: Do We Even Need Domain-Independence Anymore? [87.71321254733384]
Large language models (LLMs) can generate planning approaches tailored to specific planning problems.
LLMs can achieve state-of-the-art performance on some standard IPC domains.
We discuss whether these results signify a paradigm shift and how they can complement existing planning approaches.
arXiv Detail & Related papers (2025-01-30T22:21:12Z) - Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - Online POMDP Planning with Anytime Deterministic Guarantees [11.157761902108692]
Planning under uncertainty can be mathematically formalized using partially observable Markov decision processes (POMDPs)
Finding an optimal plan for POMDPs can be computationally expensive and is feasible only for small tasks.
We derive a deterministic relationship between a simplified solution that is easier to obtain and the theoretically optimal one.
arXiv Detail & Related papers (2023-10-03T04:40:38Z) - Lifted Sequential Planning with Lazy Constraint Generation Solvers [28.405198103927955]
This paper studies the possibilities made open by the use of Lazy Clause Generation (LCG) based approaches to Constraint Programming (CP)
We propose a novel CP model based on seminal ideas on so-called lifted causal encodings for planning as satisfiability.
We report that for planning problem instances requiring fewer plan steps our methods compare very well with the state-of-the-art in optimal sequential planning.
arXiv Detail & Related papers (2023-07-17T04:54:58Z) - Differentiable Spatial Planning using Transformers [87.90709874369192]
We propose Spatial Planning Transformers (SPT), which given an obstacle map learns to generate actions by planning over long-range spatial dependencies.
In the setting where the ground truth map is not known to the agent, we leverage pre-trained SPTs in an end-to-end framework.
SPTs outperform prior state-of-the-art differentiable planners across all the setups for both manipulation and navigation tasks.
arXiv Detail & Related papers (2021-12-02T06:48:16Z) - Safe Learning of Lifted Action Models [46.65973550325976]
We propose an algorithm for solving the model-free planning problem in classical planning.
The number of trajectories needed to solve future problems with high probability is linear in the potential size of the domain model.
arXiv Detail & Related papers (2021-07-09T01:24:01Z) - Adaptive Belief Discretization for POMDP Planning [7.508023795800546]
Many POMDP solvers uniformly discretize the belief space and give the planning error in terms of the (typically unknown) covering number.
We propose an adaptive belief discretization scheme, and give its associated planning error.
We demonstrate that our algorithm is highly competitive with the state of the art in a variety of scenarios.
arXiv Detail & Related papers (2021-04-15T07:04:32Z) - Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning [78.65083326918351]
We consider alternatives to an implicit sequential planning assumption.
We propose Divide-and-Conquer Monte Carlo Tree Search (DC-MCTS) for approximating the optimal plan.
We show that this algorithmic flexibility over planning order leads to improved results in navigation tasks in grid-worlds.
arXiv Detail & Related papers (2020-04-23T18:08:58Z) - STRIPS Action Discovery [67.73368413278631]
Recent approaches have shown the success of classical planning at synthesizing action models even when all intermediate states are missing.
We propose a new algorithm to unsupervisedly synthesize STRIPS action models with a classical planner when action signatures are unknown.
arXiv Detail & Related papers (2020-01-30T17:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.