Sample Complexity of Learning Heuristic Functions for Greedy-Best-First
and A* Search
- URL: http://arxiv.org/abs/2205.09963v3
- Date: Tue, 24 May 2022 01:31:01 GMT
- Title: Sample Complexity of Learning Heuristic Functions for Greedy-Best-First
and A* Search
- Authors: Shinsaku Sakaue, Taihei Oki
- Abstract summary: Greedy best-first search (GBFS) and A* search (A*) are popular algorithms for path-finding on large graphs.
We study the sample complexity of learning functions for GBFS and A*.
We show that our bounds for GBFS and A* under the integer-weight condition are tight up to a $lg n$ factor.
- Score: 15.191184049312467
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Greedy best-first search (GBFS) and A* search (A*) are popular algorithms for
path-finding on large graphs. Both use so-called heuristic functions, which
estimate how close a vertex is to the goal. While heuristic functions have been
handcrafted using domain knowledge, recent studies demonstrate that learning
heuristic functions from data is effective in many applications. Motivated by
this emerging approach, we study the sample complexity of learning heuristic
functions for GBFS and A*. We build on a recent framework called
\textit{data-driven algorithm design} and evaluate the
\textit{pseudo-dimension} of a class of utility functions that measure the
performance of parameterized algorithms. Assuming that a vertex set of size $n$
is fixed, we present $\mathrm{O}(n\lg n)$ and $\mathrm{O}(n^2\lg n)$ upper
bounds on the pseudo-dimensions for GBFS and A*, respectively, parameterized by
heuristic function values. The upper bound for A* can be improved to
$\mathrm{O}(n^2\lg d)$ if every vertex has a degree of at most $d$ and to
$\mathrm{O}(n \lg n)$ if edge weights are integers bounded by
$\mathrm{poly}(n)$. We also give $\Omega(n)$ lower bounds for GBFS and A*,
which imply that our bounds for GBFS and A* under the integer-weight condition
are tight up to a $\lg n$ factor. Finally, we discuss a case where the
performance of A* is measured by the suboptimality and show that we can
sometimes obtain a better guarantee by combining a parameter-dependent
worst-case bound with a sample complexity bound.
Related papers
- Two-Timescale Gradient Descent Ascent Algorithms for Nonconvex Minimax Optimization [77.3396841985172]
We provide a unified analysis of two-timescale gradient ascent (TTGDA) for solving structured non minimax optimization problems.
Our contribution is to design TTGDA algorithms are effective beyond the setting.
arXiv Detail & Related papers (2024-08-21T20:14:54Z) - On the instance optimality of detecting collisions and subgraphs [7.055459105099637]
We investigate the unlabeled instance optimality of substructure detection problems in graphs and functions.
A problem is $g(n)$-instance optimal if it admits an algorithm $A$ satisfying that for any possible input.
Our results point to a trichotomy of unlabeled instance optimality among substructure detection problems in graphs and functions.
arXiv Detail & Related papers (2023-12-15T20:50:03Z) - ReSQueing Parallel and Private Stochastic Convex Optimization [59.53297063174519]
We introduce a new tool for BFG convex optimization (SCO): a Reweighted Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density.
We develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings.
arXiv Detail & Related papers (2023-01-01T18:51:29Z) - Lower Bounds on the Worst-Case Complexity of Efficient Global
Optimization [11.523746174066702]
We derive a unified lower bound for the complexity of efficient global optimization in terms of the metric entropy of a ball in its corresponding reproducing kernel Hilbert space(RKHS)
We show that this lower bound nearly matches the upper bound attained by non-adaptive search algorithms for the commonly used squared exponential kernel and the Mat'ern kernel.
arXiv Detail & Related papers (2022-09-20T11:57:13Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Hybrid Stochastic-Deterministic Minibatch Proximal Gradient:
Less-Than-Single-Pass Optimization with Nearly Optimal Generalization [83.80460802169999]
We show that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
For loss factors, we prove that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
arXiv Detail & Related papers (2020-09-18T02:18:44Z) - Exact Quantum Query Algorithms Outperforming Parity -- Beyond The
Symmetric functions [3.652509571098291]
We first obtain optimal exact quantum query algorithms ($Q_algo(f)$) for a direct sum based class of $Omega left( 2fracsqrtn2 right)$ non-symmetric functions.
We show that query complexity of $Q_algo$ is $lceil frac3n4 rceil$ whereas $D_oplus(f)$ varies between $n-1$ and $lceil frac3n4 rce
arXiv Detail & Related papers (2020-08-14T12:17:48Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Provable guarantees for decision tree induction: the agnostic setting [16.784355746717562]
We give strengthened provable guarantees on the performance of widely employed and empirically successful sl top-down decision tree learnings.
We show that for all monotone functions$f$ and parameters $sin mathN$, theses construct a decision tree of size $stildeO((log s)/varepsilon2)$ that achieves error.
We complement our algorithmic guarantee with a near-matching $stildeOmega(log s)$ lower bound.
arXiv Detail & Related papers (2020-06-01T06:44:07Z) - Agnostic Q-learning with Function Approximation in Deterministic
Systems: Tight Bounds on Approximation Error and Sample Complexity [94.37110094442136]
We study the problem of agnostic $Q$-learning with function approximation in deterministic systems.
We show that if $delta = Oleft(rho/sqrtdim_Eright)$, then one can find the optimal policy using $Oleft(dim_Eright)$.
arXiv Detail & Related papers (2020-02-17T18:41:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.