Scaling up ML-based Black-box Planning with Partial STRIPS Models
- URL: http://arxiv.org/abs/2207.04479v1
- Date: Sun, 10 Jul 2022 14:55:16 GMT
- Title: Scaling up ML-based Black-box Planning with Partial STRIPS Models
- Authors: Matias Greco, \'Alvaro Torralba, Jorge A. Baier, Hector Palacios
- Abstract summary: We consider how a practitioner can improve ML-based black-box planning on settings where a complete symbolic model is not available.
We show that specifying an incomplete STRIPS model that describes only part of the problem enables the use of relaxations.
- Score: 3.770376172053632
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: A popular approach for sequential decision-making is to perform
simulator-based search guided with Machine Learning (ML) methods like policy
learning. On the other hand, model-relaxation heuristics can guide the search
effectively if a full declarative model is available. In this work, we consider
how a practitioner can improve ML-based black-box planning on settings where a
complete symbolic model is not available. We show that specifying an incomplete
STRIPS model that describes only part of the problem enables the use of
relaxation heuristics. Our findings on several planning domains suggest that
this is an effective way to improve ML-based black-box planning beyond
collecting more data or tuning ML architectures.
Related papers
- Exploring and Benchmarking the Planning Capabilities of Large Language Models [57.23454975238014]
This work lays the foundations for improving planning capabilities of large language models (LLMs)
We construct a comprehensive benchmark suite encompassing both classical planning benchmarks and natural language scenarios.
We investigate the use of many-shot in-context learning to enhance LLM planning, exploring the relationship between increased context length and improved planning performance.
arXiv Detail & Related papers (2024-06-18T22:57:06Z) - Black-Box Tuning of Vision-Language Models with Effective Gradient
Approximation [71.21346469382821]
We introduce collaborative black-box tuning (CBBT) for both textual prompt optimization and output feature adaptation for black-box models.
CBBT is extensively evaluated on eleven downstream benchmarks and achieves remarkable improvements compared to existing black-box VL adaptation methods.
arXiv Detail & Related papers (2023-12-26T06:31:28Z) - Interpretable Medical Diagnostics with Structured Data Extraction by
Large Language Models [59.89454513692417]
Tabular data is often hidden in text, particularly in medical diagnostic reports.
We propose a novel, simple, and effective methodology for extracting structured tabular data from textual medical reports, called TEMED-LLM.
We demonstrate that our approach significantly outperforms state-of-the-art text classification models in medical diagnostics.
arXiv Detail & Related papers (2023-06-08T09:12:28Z) - Model-based Reinforcement Learning with Multi-step Plan Value Estimation [4.158979444110977]
We introduce multi-step plans to replace multi-step actions for model-based RL.
The new model-based reinforcement learning algorithm MPPVE shows a better utilization of the learned model and achieves a better sample efficiency than state-of-the-art model-based RL approaches.
arXiv Detail & Related papers (2022-09-12T18:22:11Z) - Model-Agnostic Multitask Fine-tuning for Few-shot Vision-Language
Transfer Learning [59.38343286807997]
We propose Model-Agnostic Multitask Fine-tuning (MAMF) for vision-language models on unseen tasks.
Compared with model-agnostic meta-learning (MAML), MAMF discards the bi-level optimization and uses only first-order gradients.
We show that MAMF consistently outperforms the classical fine-tuning method for few-shot transfer learning on five benchmark datasets.
arXiv Detail & Related papers (2022-03-09T17:26:53Z) - Evaluating model-based planning and planner amortization for continuous
control [79.49319308600228]
We take a hybrid approach, combining model predictive control (MPC) with a learned model and model-free policy learning.
We find that well-tuned model-free agents are strong baselines even for high DoF control problems.
We show that it is possible to distil a model-based planner into a policy that amortizes the planning without any loss of performance.
arXiv Detail & Related papers (2021-10-07T12:00:40Z) - MeLIME: Meaningful Local Explanation for Machine Learning Models [2.819725769698229]
We show that our approach, MeLIME, produces more meaningful explanations compared to other techniques over different ML models.
MeLIME generalizes the LIME method, allowing more flexible perturbation sampling and the use of different local interpretable models.
arXiv Detail & Related papers (2020-09-12T16:06:58Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - Policy-Aware Model Learning for Policy Gradient Methods [29.129883702165774]
This paper considers the problem of learning a model in model-based reinforcement learning (MBRL)
We propose that the model learning module should incorporate the way the planner is going to use the model.
We call this approach Policy-Aware Model Learning (PAML)
arXiv Detail & Related papers (2020-02-28T19:18:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.