A* Search Without Expansions: Learning Heuristic Functions with Deep
Q-Networks
- URL: http://arxiv.org/abs/2102.04518v2
- Date: Thu, 23 Mar 2023 17:38:09 GMT
- Title: A* Search Without Expansions: Learning Heuristic Functions with Deep
Q-Networks
- Authors: Forest Agostinelli, Alexander Shmakov, Stephen McAleer, Roy Fox,
Pierre Baldi
- Abstract summary: We introduce Q* search, a search algorithm that uses deep Q-networks to guide search in order to take advantage of the sum of the transition costs and values of the children of a node.
We use Q* search to solve the Rubik's cube when formulated with a large action space that includes 1872 meta-actions.
Q* search is up to 129 times faster and generates up to 1288 times fewer nodes than A* search.
- Score: 70.0197695666261
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Efficiently solving problems with large action spaces using A* search has
been of importance to the artificial intelligence community for decades. This
is because the computation and memory requirements of A* search grow linearly
with the size of the action space. This burden becomes even more apparent when
A* search uses a heuristic function learned by computationally expensive
function approximators, such as deep neural networks. To address this problem,
we introduce Q* search, a search algorithm that uses deep Q-networks to guide
search in order to take advantage of the fact that the sum of the transition
costs and heuristic values of the children of a node can be computed with a
single forward pass through a deep Q-network without explicitly generating
those children. This significantly reduces computation time and requires only
one node to be generated per iteration. We use Q* search to solve the Rubik's
cube when formulated with a large action space that includes 1872 meta-actions
and find that this 157-fold increase in the size of the action space incurs
less than a 4-fold increase in computation time and less than a 3-fold increase
in number of nodes generated when performing Q* search. Furthermore, Q* search
is up to 129 times faster and generates up to 1288 times fewer nodes than A*
search. Finally, although obtaining admissible heuristic functions from deep
neural networks is an ongoing area of research, we prove that Q* search is
guaranteed to find a shortest path given a heuristic function that neither
overestimates the cost of a shortest path nor underestimates the transition
cost.
Related papers
- LiteSearch: Efficacious Tree Search for LLM [70.29796112457662]
This study introduces a novel guided tree search algorithm with dynamic node selection and node-level exploration budget.
Experiments conducted on the GSM8K and TabMWP datasets demonstrate that our approach enjoys significantly lower computational costs compared to baseline methods.
arXiv Detail & Related papers (2024-06-29T05:14:04Z) - SLOPE: Search with Learned Optimal Pruning-based Expansion [2.0618817976970103]
We present the Search with Learned Optimal Pruning-based Expansion (SLOPE)
It learns the distance of a node from a possible optimal path, which in turn reduces the size of the open list.
This ensures that the search explores only the region close to optimal paths while lowering memory and computational costs.
arXiv Detail & Related papers (2024-06-07T13:42:15Z) - Autonomous Tree-search Ability of Large Language Models [58.68735916408101]
Large Language Models have excelled in remarkable reasoning capabilities with advanced prompting techniques.
Recent works propose to utilize external programs to define search logic, such that LLMs can perform passive tree search to solve more challenging reasoning tasks.
We propose a new concept called autonomous tree-search ability of LLM, which can automatically generate a response containing search trajectories for the correct answer.
arXiv Detail & Related papers (2023-10-14T14:14:38Z) - Learning Graph Search Heuristics [48.83557172525969]
We present PHIL (Path Heuristic with Imitation Learning), a novel neural architecture and a training algorithm for discovering graph search and navigations from data.
Our function learns graph embeddings useful for inferring node distances, runs in constant time independent of graph sizes, and can be easily incorporated in an algorithm such as A* at test time.
Experiments show that PHIL reduces the number of explored nodes compared to state-of-the-art methods on benchmark datasets by 58.5% on average.
arXiv Detail & Related papers (2022-12-07T22:28:00Z) - A Differentiable Loss Function for Learning Heuristics in A* [0.0]
This paper argues that this does not necessarily lead to a faster search of A* algorithm since its execution relies on relative values instead of absolute ones.
As a mitigation, we propose a L* loss, which upper-bounds the number of excessively expanded states inside the A* search.
The L* loss, when used in the optimization of state-of-the-art deep neural networks for automated planning in maze domains like Sokoban and maze with teleports, significantly improves the fraction of solved problems, the quality of founded plans, and reduces the number of expanded states to approximately 50%.
arXiv Detail & Related papers (2022-09-12T12:43:05Z) - The (Un)Scalability of Heuristic Approximators for NP-Hard Search
Problems [25.641418039598587]
The A* algorithm is commonly used to solve NP-hard optimization problems.
We show that accurate approximation for many such problems is also NP-hard.
arXiv Detail & Related papers (2022-09-07T18:02:02Z) - Learning heuristics for A* [9.701208207491879]
We combine recent advancements in Neuralic Reasoning to learn efficient functions for path finding problems on graphs.
We exploit multi-task learning to learn Dijkstra's algorithm and a consistent function for the A* search algorithm.
Results show that running A* over the learnts value can greatly speed up target node searching compared to Dijkstra.
arXiv Detail & Related papers (2022-04-11T10:13:53Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.