Adjust Planning Strategies to Accommodate Reinforcement Learning Agents
- URL: http://arxiv.org/abs/2003.08554v1
- Date: Thu, 19 Mar 2020 03:35:10 GMT
- Title: Adjust Planning Strategies to Accommodate Reinforcement Learning Agents
- Authors: Xuerun Chen
- Abstract summary: We create an optimization strategy for planning parameters, through analysis to the connection of reaction and planning.
The whole algorithm can find a satisfactory setting of planning parameters, making full use of reaction capability of specific agents.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In agent control issues, the idea of combining reinforcement learning and
planning has attracted much attention. Two methods focus on micro and macro
action respectively. Their advantages would show together if there is a good
cooperation between them. An essential for the cooperation is to find an
appropriate boundary, assigning different functions to each method. Such
boundary could be represented by parameters in a planning algorithm. In this
paper, we create an optimization strategy for planning parameters, through
analysis to the connection of reaction and planning; we also create a
non-gradient method for accelerating the optimization. The whole algorithm can
find a satisfactory setting of planning parameters, making full use of reaction
capability of specific agents.
Related papers
- DHP: Discrete Hierarchical Planning for Hierarchical Reinforcement Learning Agents [2.1438108757511958]
Our key contribution is a Discrete Hierarchical Planning (DHP) method, an alternative to traditional distance-based approaches.
We provide theoretical foundations for the method and demonstrate its effectiveness through extensive empirical evaluations.
We evaluate our method on long-horizon visual planning tasks in a 25-room environment, where it significantly outperforms previous benchmarks at success rate and average episode length.
arXiv Detail & Related papers (2025-02-04T03:05:55Z) - Parallel Strategies for Best-First Generalized Planning [51.713634067802104]
Generalized planning (GP) is a research area of AI that studies the automated synthesis of algorithmic-like solutions capable of solving multiple classical planning instances.
One of the current advancements has been the introduction of Best-First Generalized Planning (BFGP), a GP algorithm based on a novel solution space that can be explored with search.
This paper evaluates the application of parallel search techniques to BFGP, another critical component in closing the performance gap.
arXiv Detail & Related papers (2024-07-31T09:50:22Z) - Cooperative Bayesian Optimization for Imperfect Agents [32.15315995944448]
Two agents choose together at which points to query the function but have only control over one variable each.
We formulate the solution as sequential decision-making, where the agent we control models the user as a computationally rational agent with prior knowledge about the function.
We show that strategic planning of the queries enables better identification of the global maximum of the function as long as the user avoids excessive exploration.
arXiv Detail & Related papers (2024-03-07T12:16:51Z) - Experiment Planning with Function Approximation [49.50254688629728]
We study the problem of experiment planning with function approximation in contextual bandit problems.
We propose two experiment planning strategies compatible with function approximation.
We show that a uniform sampler achieves competitive optimality rates in the setting where the number of actions is small.
arXiv Detail & Related papers (2024-01-10T14:40:23Z) - Planning as In-Painting: A Diffusion-Based Embodied Task Planning
Framework for Environments under Uncertainty [56.30846158280031]
Task planning for embodied AI has been one of the most challenging problems.
We propose a task-agnostic method named 'planning as in-painting'
The proposed framework achieves promising performances in various embodied AI tasks.
arXiv Detail & Related papers (2023-12-02T10:07:17Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Optimal Cost-Preference Trade-off Planning with Multiple Temporal Tasks [3.655021726150368]
We introduce a novel notion of preference that provides a generalized framework to express preferences over individual tasks as well as their relations.
We perform an optimal trade-off (Pareto) analysis between behaviors that adhere to the user's preference and the ones that are resource optimal.
arXiv Detail & Related papers (2023-06-22T21:56:49Z) - Efficient Planning in Combinatorial Action Spaces with Applications to
Cooperative Multi-Agent Reinforcement Learning [16.844525262228103]
In cooperative multi-agent reinforcement learning, a potentially large number of agents jointly optimize a global reward function, which leads to a blow-up in the action space by the number of agents.
As a minimal requirement, we assume access to an argmax oracle that allows to efficiently compute the greedy policy for any Q-function in the model class.
We propose efficient algorithms for this setting that lead to compute and query complexity in all relevant problem parameters.
arXiv Detail & Related papers (2023-02-08T23:42:49Z) - Learn to Match with No Regret: Reinforcement Learning in Markov Matching
Markets [151.03738099494765]
We study a Markov matching market involving a planner and a set of strategic agents on the two sides of the market.
We propose a reinforcement learning framework that integrates optimistic value iteration with maximum weight matching.
We prove that the algorithm achieves sublinear regret.
arXiv Detail & Related papers (2022-03-07T19:51:25Z) - Extended Task and Motion Planning of Long-horizon Robot Manipulation [28.951816622135922]
Task and Motion Planning (TAMP) requires integration of symbolic reasoning with metric motion planning.
Most TAMP approaches fail to provide feasible solutions when there is missing knowledge about the environment at the symbolic level.
We propose a novel approach for decision-making on extended decision spaces over plan skeletons and action parameters.
arXiv Detail & Related papers (2021-03-09T14:44:08Z) - A Unifying Framework for Reinforcement Learning and Planning [2.564530030795554]
This paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP)
At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions.
arXiv Detail & Related papers (2020-06-26T14:30:41Z) - Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning [78.65083326918351]
We consider alternatives to an implicit sequential planning assumption.
We propose Divide-and-Conquer Monte Carlo Tree Search (DC-MCTS) for approximating the optimal plan.
We show that this algorithmic flexibility over planning order leads to improved results in navigation tasks in grid-worlds.
arXiv Detail & Related papers (2020-04-23T18:08:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.