Safe-Planner: A Single-Outcome Replanner for Computing Strong Cyclic
Policies in Fully Observable Non-Deterministic Domains
- URL: http://arxiv.org/abs/2109.11471v1
- Date: Thu, 23 Sep 2021 16:20:35 GMT
- Title: Safe-Planner: A Single-Outcome Replanner for Computing Strong Cyclic
Policies in Fully Observable Non-Deterministic Domains
- Authors: Vahid Mokhtari, Ajay Suresha Sathya, Nikolaos Tsiogkas, Wilm Decre
- Abstract summary: We introduce an offline replanner, called Safe-Planner, that relies on a single-outcome determinization to compile a non-deterministic domain to a set of classical domains.
We show experimentally that this approach can allow SP to avoid generating misleading plans but to generate weak plans that directly lead to strong solutions.
- Score: 0.22940141855172028
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Replanners are efficient methods for solving non-deterministic planning
problems. Despite showing good scalability, existing replanners often fail to
solve problems involving a large number of misleading plans, i.e., weak plans
that do not lead to strong solutions, however, due to their minimal lengths,
are likely to be found at every replanning iteration. The poor performance of
replanners in such problems is due to their all-outcome determinization. That
is, when compiling from non-deterministic to classical, they include all
compiled classical operators in a single deterministic domain which leads
replanners to continually generate misleading plans. We introduce an offline
replanner, called Safe-Planner (SP), that relies on a single-outcome
determinization to compile a non-deterministic domain to a set of classical
domains, and ordering heuristics for ranking the obtained classical domains.
The proposed single-outcome determinization and the heuristics allow for
alternating between different classical domains. We show experimentally that
this approach can allow SP to avoid generating misleading plans but to generate
weak plans that directly lead to strong solutions. The experiments show that SP
outperforms state-of-the-art non-deterministic solvers by solving a broader
range of problems. We also validate the practical utility of SP in real-world
non-deterministic robotic tasks.
Related papers
- Parallel Strategies for Best-First Generalized Planning [51.713634067802104]
Generalized planning (GP) is a research area of AI that studies the automated synthesis of algorithmic-like solutions capable of solving multiple classical planning instances.
One of the current advancements has been the introduction of Best-First Generalized Planning (BFGP), a GP algorithm based on a novel solution space that can be explored with search.
This paper evaluates the application of parallel search techniques to BFGP, another critical component in closing the performance gap.
arXiv Detail & Related papers (2024-07-31T09:50:22Z) - Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - LLM-Assist: Enhancing Closed-Loop Planning with Language-Based Reasoning [65.86754998249224]
We develop a novel hybrid planner that leverages a conventional rule-based planner in conjunction with an LLM-based planner.
Our approach navigates complex scenarios which existing planners struggle with, produces well-reasoned outputs while also remaining grounded through working alongside the rule-based approach.
arXiv Detail & Related papers (2023-12-30T02:53:45Z) - Online POMDP Planning with Anytime Deterministic Guarantees [11.157761902108692]
Planning under uncertainty can be mathematically formalized using partially observable Markov decision processes (POMDPs)
Finding an optimal plan for POMDPs can be computationally expensive and is feasible only for small tasks.
We derive a deterministic relationship between a simplified solution that is easier to obtain and the theoretically optimal one.
arXiv Detail & Related papers (2023-10-03T04:40:38Z) - Lifted Sequential Planning with Lazy Constraint Generation Solvers [28.405198103927955]
This paper studies the possibilities made open by the use of Lazy Clause Generation (LCG) based approaches to Constraint Programming (CP)
We propose a novel CP model based on seminal ideas on so-called lifted causal encodings for planning as satisfiability.
We report that for planning problem instances requiring fewer plan steps our methods compare very well with the state-of-the-art in optimal sequential planning.
arXiv Detail & Related papers (2023-07-17T04:54:58Z) - Temporal Planning with Incomplete Knowledge and Perceptual Information [0.0]
This paper presents a new planning approach that combines contingent plan construction within a temporal planning framework.
We propose a small extension to the Planning Domain Definition Language (PDDL) to model incomplete, (ii) knowledge sensing actions.
We also introduce a new set of planning domains to evaluate our solver, which has shown good performance on a variety of problems.
arXiv Detail & Related papers (2022-07-20T07:26:08Z) - Differentiable Spatial Planning using Transformers [87.90709874369192]
We propose Spatial Planning Transformers (SPT), which given an obstacle map learns to generate actions by planning over long-range spatial dependencies.
In the setting where the ground truth map is not known to the agent, we leverage pre-trained SPTs in an end-to-end framework.
SPTs outperform prior state-of-the-art differentiable planners across all the setups for both manipulation and navigation tasks.
arXiv Detail & Related papers (2021-12-02T06:48:16Z) - Safe Learning of Lifted Action Models [46.65973550325976]
We propose an algorithm for solving the model-free planning problem in classical planning.
The number of trajectories needed to solve future problems with high probability is linear in the potential size of the domain model.
arXiv Detail & Related papers (2021-07-09T01:24:01Z) - Adaptive Belief Discretization for POMDP Planning [7.508023795800546]
Many POMDP solvers uniformly discretize the belief space and give the planning error in terms of the (typically unknown) covering number.
We propose an adaptive belief discretization scheme, and give its associated planning error.
We demonstrate that our algorithm is highly competitive with the state of the art in a variety of scenarios.
arXiv Detail & Related papers (2021-04-15T07:04:32Z) - Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning [78.65083326918351]
We consider alternatives to an implicit sequential planning assumption.
We propose Divide-and-Conquer Monte Carlo Tree Search (DC-MCTS) for approximating the optimal plan.
We show that this algorithmic flexibility over planning order leads to improved results in navigation tasks in grid-worlds.
arXiv Detail & Related papers (2020-04-23T18:08:58Z) - STRIPS Action Discovery [67.73368413278631]
Recent approaches have shown the success of classical planning at synthesizing action models even when all intermediate states are missing.
We propose a new algorithm to unsupervisedly synthesize STRIPS action models with a classical planner when action signatures are unknown.
arXiv Detail & Related papers (2020-01-30T17:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.