HDDL 2.1: Towards Defining a Formalism and a Semantics for Temporal HTN
Planning
- URL: http://arxiv.org/abs/2306.07353v1
- Date: Mon, 12 Jun 2023 18:21:23 GMT
- Title: HDDL 2.1: Towards Defining a Formalism and a Semantics for Temporal HTN
Planning
- Authors: Damien Pellier, Alexandre Albore, Humbert Fiorino, Rafael Bailon-Ruiz
- Abstract summary: Real world applications need modelling rich and diverse automated planning problems.
hierarchical task network (HTN) formalism does not allow to represent planning problems with numerical and temporal constraints.
We propose to fill the gap between HDDL and these operational needs and to extend HDDL by taking inspiration from PDDL 2.1.
- Score: 64.07762708909846
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real world applications as in industry and robotics need modelling rich and
diverse automated planning problems. Their resolution usually requires
coordinated and concurrent action execution. In several cases, these problems
are naturally decomposed in a hierarchical way and expressed by a Hierarchical
Task Network (HTN) formalism.
HDDL, a hierarchical extension of the Planning Domain Definition Language
(PDDL), unlike PDDL 2.1 does not allow to represent planning problems with
numerical and temporal constraints, which are essential for real world
applications. We propose to fill the gap between HDDL and these operational
needs and to extend HDDL by taking inspiration from PDDL 2.1 in order to
express numerical and temporal expressions. This paper opens discussions on the
semantics and the syntax needed for a future HDDL 2.1 extension.
Related papers
- Generating Symbolic World Models via Test-time Scaling of Large Language Models [28.258707611580643]
Planning Domain Definition Language (PDDL) is leveraged as a planning abstraction that enables precise and formal state descriptions.
We introduce a simple yet effective algorithm, which first employs a Best-of-N sampling approach to improve the quality of the initial solution and then refines the solution in a fine-grained manner with verbalized machine learning.
Our method outperforms o1-mini by a considerable margin in the generation of PDDL domain, achieving over 50% success rate on two tasks.
arXiv Detail & Related papers (2025-02-07T07:52:25Z) - Planning with Vision-Language Models and a Use Case in Robot-Assisted Teaching [0.9217021281095907]
This paper introduces Image2PDDL, a novel framework that leverages Vision-Language Models (VLMs) to automatically convert images of initial states and descriptions of goal states into PDDL problems.
We evaluate the framework on various domains, including standard planning domains like blocksworld and sliding tile puzzles, using datasets with multiple difficulty levels.
We will discuss a potential use case in robot-assisted teaching of students with Autism Spectrum Disorder.
arXiv Detail & Related papers (2025-01-29T14:04:54Z) - SCoTT: Wireless-Aware Path Planning with Vision Language Models and Strategic Chains-of-Thought [78.53885607559958]
A novel approach using vision language models (VLMs) is proposed for enabling path planning in complex wireless-aware environments.
To this end, insights from a digital twin with real-world wireless ray tracing data are explored.
Results show that SCoTT achieves very close average path gains compared to DP-WA* while at the same time yielding consistently shorter path lengths.
arXiv Detail & Related papers (2024-11-27T10:45:49Z) - Leveraging Environment Interaction for Automated PDDL Translation and Planning with Large Language Models [7.3238629831871735]
Large Language Models (LLMs) have shown remarkable performance in various natural language tasks.
Planning problems into the Planning Domain Definition Language (PDDL) has been proposed as a potential solution.
We propose a novel approach that leverages LLMs and environment feedback to automatically generate PDDL domain and problem description files.
arXiv Detail & Related papers (2024-07-17T19:50:51Z) - PROC2PDDL: Open-Domain Planning Representations from Texts [56.627183903841164]
Proc2PDDL is the first dataset containing open-domain procedural texts paired with expert-annotated PDDL representations.
We show that Proc2PDDL is highly challenging, with GPT-3.5's success rate close to 0% and GPT-4's around 35%.
arXiv Detail & Related papers (2024-02-29T19:40:25Z) - Real-World Planning with PDDL+ and Beyond [55.73913765642435]
We present Nyx, a novel PDDL+ planner built to emphasize lightness, simplicity, and, most importantly, adaptability.
Nyx can be tailored to virtually any potential real-world application requiring some form of AI Planning, paving the way for wider adoption of planning methods for solving real-world problems.
arXiv Detail & Related papers (2024-02-19T07:35:49Z) - Serving Deep Learning Model in Relational Databases [70.53282490832189]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-art DL-centric architecture offloads DL computations to dedicated DL frameworks.
The potential UDF-centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the relational database management system (RDBMS)
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - HDDL 2.1: Towards Defining an HTN Formalism with Time [0.0]
Real world applications of planning, like in industry and robotics, require modelling rich and diverse scenarios.
Their resolution usually requires coordinated and concurrent action executions.
In several cases, such planning problems are naturally decomposed in a hierarchical way and expressed by a Hierarchical Task Network formalism.
This paper opens discussions on the semantics and the syntax needed to extend HDDL, and illustrate these needs with the modelling of an Earth Observing Satellite planning problem.
arXiv Detail & Related papers (2022-06-03T21:22:19Z) - Tensor Processing Primitives: A Programming Abstraction for Efficiency
and Portability in Deep Learning Workloads [86.62083829086393]
This work introduces the Processing Primitives (TPP), a programming abstraction striving for efficient, portable implementation of Deep Learning-workloads with high-productivity.
TPPs define a compact, yet versatile set of 2D-tensor operators (or a virtual ISA), which can be utilized as building-blocks to construct complex operators on high-dimensional tensors.
We demonstrate the efficacy of our approach using standalone kernels and end-to-end DL-workloads expressed entirely via TPPs that outperform state-of-the-art implementations on multiple platforms.
arXiv Detail & Related papers (2021-04-12T18:35:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.