AMLSI: A Novel Accurate Action Model Learning Algorithm
- URL: http://arxiv.org/abs/2011.13277v1
- Date: Thu, 26 Nov 2020 13:25:08 GMT
- Title: AMLSI: A Novel Accurate Action Model Learning Algorithm
- Authors: Maxence Grand, Humbert Fiorino, Damien Pellier
- Abstract summary: The AMLSI approach does not require a training dataset of plan traces to work.
AMLSI proceeds by trial and error: it queries the system to learn with randomly generated action sequences.
Unlike other algorithms, we show that AMLSI is able to lift this lock by learning domains from partial and noisy observations.
- Score: 1.1797787239802762
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper presents new approach based on grammar induction called AMLSI
Action Model Learning with State machine Interactions. The AMLSI approach does
not require a training dataset of plan traces to work. AMLSI proceeds by trial
and error: it queries the system to learn with randomly generated action
sequences, and it observes the state transitions of the system, then AMLSI
returns a PDDL domain corresponding to the system. A key issue for domain
learning is the ability to plan with the learned domains. It often happens that
a small learning error leads to a domain that is unusable for planning. Unlike
other algorithms, we show that AMLSI is able to lift this lock by learning
domains from partial and noisy observations with sufficient accuracy to allow
planners to solve new problems.
Related papers
- Deep Learning for Generalised Planning with Background Knowledge [0.0]
Planning problems are easy to solve but hard to optimise.
We propose a new machine learning approach that allows users to specify background knowledge.
By incorporating BK, our approach bypasses the need to relearn how to solve problems from scratch and instead focuses the learning on plan quality optimisation.
arXiv Detail & Related papers (2024-10-10T13:49:05Z) - Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models [79.28821338925947]
Domain-Class Incremental Learning is a realistic but challenging continual learning scenario.
To handle these diverse tasks, pre-trained Vision-Language Models (VLMs) are introduced for their strong generalizability.
This incurs a new problem: the knowledge encoded in the pre-trained VLMs may be disturbed when adapting to new tasks, compromising their inherent zero-shot ability.
Existing methods tackle it by tuning VLMs with knowledge distillation on extra datasets, which demands heavy overhead.
We propose the Distribution-aware Interference-free Knowledge Integration (DIKI) framework, retaining pre-trained knowledge of
arXiv Detail & Related papers (2024-07-07T12:19:37Z) - A New View on Planning in Online Reinforcement Learning [19.35031543927374]
This paper investigates a new approach to model-based reinforcement learning using background planning.
We show that our GSP algorithm can propagate value from an abstract space in a manner that helps a variety of base learners learn significantly faster in different domains.
arXiv Detail & Related papers (2024-06-03T17:45:19Z) - Learning the Finer Things: Bayesian Structure Learning at the
Instantiation Level [0.0]
Successful machine learning methods require a trade-off between memorization and generalization.
We present a novel probabilistic graphical model structure learning approach that can learn, generalize and explain in elusive domains.
arXiv Detail & Related papers (2023-03-08T02:31:49Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - Domain Adaptation Broad Learning System Based on Locally Linear
Embedding [3.274290296343038]
The proposed algorithm can learn a robust classification model by using a small part of labeled data from the target domain and all labeled data from the source domain.
The results show that our approach can get better classification accuracy with less running time than many existing transfer learning approaches.
arXiv Detail & Related papers (2021-06-28T01:55:57Z) - How Fine-Tuning Allows for Effective Meta-Learning [50.17896588738377]
We present a theoretical framework for analyzing representations derived from a MAML-like algorithm.
We provide risk bounds on the best predictor found by fine-tuning via gradient descent, demonstrating that the algorithm can provably leverage the shared structure.
This separation result underscores the benefit of fine-tuning-based methods, such as MAML, over methods with "frozen representation" objectives in few-shot learning.
arXiv Detail & Related papers (2021-05-05T17:56:00Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - STRIPS Action Discovery [67.73368413278631]
Recent approaches have shown the success of classical planning at synthesizing action models even when all intermediate states are missing.
We propose a new algorithm to unsupervisedly synthesize STRIPS action models with a classical planner when action signatures are unknown.
arXiv Detail & Related papers (2020-01-30T17:08:39Z) - Hierarchical Variational Imitation Learning of Control Programs [131.7671843857375]
We propose a variational inference method for imitation learning of a control policy represented by parametrized hierarchical procedures (PHP)
Our method discovers the hierarchical structure in a dataset of observation-action traces of teacher demonstrations, by learning an approximate posterior distribution over the latent sequence of procedure calls and terminations.
We demonstrate a novel benefit of variational inference in the context of hierarchical imitation learning: in decomposing the policy into simpler procedures, inference can leverage acausal information that is unused by other methods.
arXiv Detail & Related papers (2019-12-29T08:57:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.