Removing Planner Bias in Goal Recognition Through Multi-Plan Dataset Generation
- URL: http://arxiv.org/abs/2602.14691v1
- Date: Mon, 16 Feb 2026 12:25:35 GMT
- Title: Removing Planner Bias in Goal Recognition Through Multi-Plan Dataset Generation
- Authors: Mustafa F. Abdelwahed, Felipe Meneguzzi Kin Max Piamolini Gusmao, Joan Espasa,
- Abstract summary: All existing datasets suffer from a systematical bias induced by the planning systems that generated them.<n>We propose a new method that uses top-k planning to generate multiple, different, plans for the same goal hypothesis.<n>This allows us to introduce a new metric called Version Coverage Score (VCS) to measure the resilience of the goal recogniser when inferring a goal based on different sets of plans.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autonomous agents require some form of goal and plan recognition to interact in multiagent settings. Unfortunately, all existing goal recognition datasets suffer from a systematical bias induced by the planning systems that generated them, namely heuristic-based forward search. This means that existing datasets lack enough challenge for more realistic scenarios (e.g., agents using different planners), which impacts the evaluation of goal recognisers with respect to using different planners for the same goal. In this paper, we propose a new method that uses top-k planning to generate multiple, different, plans for the same goal hypothesis, yielding benchmarks that mitigate the bias found in the current dataset. This allows us to introduce a new metric called Version Coverage Score (VCS) to measure the resilience of the goal recogniser when inferring a goal based on different sets of plans. Our results show that the resilience of the current state-of-the-art goal recogniser degrades substantially under low observability settings.
Related papers
- Locally Adaptive Multi-Objective Learning [50.29753546978998]
We work in an online setting where the data distribution can change arbitrarily over time.<n>Existing approaches to this problem aim to minimize the set of objectives over the entire time horizon.<n>We consider an alternative procedure that achieves local adaptivity by replacing one part of the multi-objective learning method with an adaptive online algorithm.
arXiv Detail & Related papers (2026-02-16T17:31:48Z) - Decentralized Multi-Agent Goal Assignment for Path Planning using Large Language Models [7.94408712915778]
This work addresses the problem of decentralized goal assignment for multi-agent path planning.<n>Agents independently generate ranked preferences over goals based on structured representations of the environment.<n>We compare greedys, optimal assignment, and large language model (LLM)-based agents in fully observable grid-world settings.
arXiv Detail & Related papers (2025-10-27T20:05:56Z) - Hierarchical Object-Oriented POMDP Planning for Object Rearrangement [19.62753215239688]
Current object rearrangement solutions, primarily based on Reinforcement Learning or hand-coded planning methods, often lack adaptability to diverse challenges.<n>To address this limitation, we introduce a novel Hierarchical Object-Oriented Partially Observed Markov Decision Process (HOO-POMDP) planning approach.<n>We present an online planning framework and a new benchmark dataset for solving multi-object rearrangement problems in partially observable, multi-room environments.
arXiv Detail & Related papers (2024-12-02T10:19:36Z) - Propose, Assess, Search: Harnessing LLMs for Goal-Oriented Planning in Instructional Videos [48.15438373870542]
VidAssist is an integrated framework designed for zero/few-shot goal-oriented planning in instructional videos.
It employs a breadth-first search algorithm for optimal plan generation.
Experiments demonstrate that VidAssist offers a unified framework for different goal-oriented planning setups.
arXiv Detail & Related papers (2024-09-30T17:57:28Z) - Temporally Extended Goal Recognition in Fully Observable
Non-Deterministic Domain Models [43.460098744623416]
Existing approaches assume that goal hypotheses comprise a single conjunctive formula over a single final state.
We focus on temporally extended goals in Fully Observable Non-Deterministic (FOND) planning domain models.
Empirical results show that our approach is accurate in recognizing temporally extended goals in different recognition settings.
arXiv Detail & Related papers (2023-06-14T18:02:00Z) - Imitating Graph-Based Planning with Goal-Conditioned Policies [72.61631088613048]
We present a self-imitation scheme which distills a subgoal-conditioned policy into the target-goal-conditioned policy.
We empirically show that our method can significantly boost the sample-efficiency of the existing goal-conditioned RL methods.
arXiv Detail & Related papers (2023-03-20T14:51:10Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - Generative multitask learning mitigates target-causing confounding [61.21582323566118]
We propose a simple and scalable approach to causal representation learning for multitask learning.
The improvement comes from mitigating unobserved confounders that cause the targets, but not the input.
Our results on the Attributes of People and Taskonomy datasets reflect the conceptual improvement in robustness to prior probability shift.
arXiv Detail & Related papers (2022-02-08T20:42:14Z) - Unsupervised and self-adaptative techniques for cross-domain person
re-identification [82.54691433502335]
Person Re-Identification (ReID) across non-overlapping cameras is a challenging task.
Unsupervised Domain Adaptation (UDA) is a promising alternative, as it performs feature-learning adaptation from a model trained on a source to a target domain without identity-label annotation.
In this paper, we propose a novel UDA-based ReID method that takes advantage of triplets of samples created by a new offline strategy.
arXiv Detail & Related papers (2021-03-21T23:58:39Z) - PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals [14.315501760755609]
PlanGAN is a model-based algorithm for solving multi-goal tasks in environments with sparse rewards.
Our studies indicate that PlanGAN can achieve comparable performance whilst being around 4-8 times more sample efficient.
arXiv Detail & Related papers (2020-06-01T12:53:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.