Testing and reconstruction via decision trees
- URL: http://arxiv.org/abs/2012.08735v1
- Date: Wed, 16 Dec 2020 04:18:00 GMT
- Title: Testing and reconstruction via decision trees
- Authors: Guy Blanc, Jane Lange, Li-Yang Tan
- Abstract summary: We study sublinear and local computation algorithms for decision trees, focusing on testing and reconstruction.
A tester that runs in $mathrmpoly(log s, 1/varepsilon)cdot nlog n$ time makes $mathrmpoly(log s,1/varepsilon)cdot n$ queries to an unknown function.
- Score: 19.304587350775385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study sublinear and local computation algorithms for decision trees,
focusing on testing and reconstruction. Our first result is a tester that runs
in $\mathrm{poly}(\log s, 1/\varepsilon)\cdot n\log n$ time, makes
$\mathrm{poly}(\log s,1/\varepsilon)\cdot \log n$ queries to an unknown
function $f$, and:
$\circ$ Accepts if $f$ is $\varepsilon$-close to a size-$s$ decision tree;
$\circ$ Rejects if $f$ is $\Omega(\varepsilon)$-far from decision trees of
size $s^{\tilde{O}((\log s)^2/\varepsilon^2)}$.
Existing testers distinguish size-$s$ decision trees from those that are
$\varepsilon$-far from from size-$s$ decision trees in
$\mathrm{poly}(s^s,1/\varepsilon)\cdot n$ time with $\tilde{O}(s/\varepsilon)$
queries. We therefore solve an incomparable problem, but achieve
doubly-exponential-in-$s$ and exponential-in-$s$ improvements in time and query
complexities respectively. We obtain our tester by designing a reconstruction
algorithm for decision trees: given query access to a function $f$ that is
close to a small decision tree, this algorithm provides fast query access to a
small decision tree that is close to $f$. By known relationships, our results
yield reconstruction algorithms for numerous other boolean function properties
-- Fourier degree, randomized and quantum query complexities, certificate
complexity, sensitivity, etc. -- which in turn yield new testers for these
properties. Finally, we give a hardness result for testing whether an unknown
function is $\varepsilon$-close-to or $\Omega(\varepsilon)$-far-from size-$s$
decision trees. We show that an efficient algorithm for this task would yield
an efficient algorithm for properly learning decision trees, a central open
problem of learning theory. It has long been known that proper learning
algorithms for any class $\mathcal{H}$ yield property testers for
$\mathcal{H}$; this provides an example of a converse.
Related papers
- Near-Optimal Bounds for Learning Gaussian Halfspaces with Random
Classification Noise [50.64137465792738]
We show that any efficient SQ algorithm for the problem requires sample complexity at least $Omega(d1/2/(maxp, epsilon)2)$.
Our lower bound suggests that this quadratic dependence on $1/epsilon$ is inherent for efficient algorithms.
arXiv Detail & Related papers (2023-07-13T18:59:28Z) - Most Neural Networks Are Almost Learnable [52.40331776572531]
We show that for any fixed $epsilon>0$ and depth $i$, there is a poly-time algorithm that learns random Xavier networks of depth $i$.
The algorithm runs in time and sample complexity of $(bard)mathrmpoly(epsilon-1)$, where $bar d$ is the size of the network.
For some cases of sigmoid and ReLU-like activations the bound can be improved to $(bard)mathrmpolylog(eps
arXiv Detail & Related papers (2023-05-25T22:27:42Z) - Fully-Dynamic Approximate Decision Trees With Worst-Case Update Time
Guarantees [3.5509551353363644]
We give the first algorithm that maintains an approximate decision tree over an arbitrary sequence of insertions and deletions of labeled examples.
We provide a deterministic algorithm that maintains an $epsilon$-approximate tree using $O!left(fracd, f(n)n operatornamepolyfrachepsilonright)$ operations per update.
arXiv Detail & Related papers (2023-02-08T11:02:58Z) - Superpolynomial Lower Bounds for Decision Tree Learning and Testing [5.117810469621253]
We prove, under randomized ETH, superpolynomial lower bounds for two basic problems.
Our results imply new lower bounds for distribution-free PAC learning and testing of decision trees.
We show our lower bound for learning decision trees can be improved to $nOmega(log s)$.
arXiv Detail & Related papers (2022-10-12T16:25:48Z) - On Learning and Testing Decision Tree [0.22843885788439797]
We give a new algorithm that achieves $phi(s,1/epsilon)=2O(log3s)/epsilon3)$.
We also study the testability of depth-$d$ decision tree and give a it distribution free tester.
arXiv Detail & Related papers (2021-08-10T11:04:06Z) - Learning stochastic decision trees [19.304587350775385]
We give a quasipolynomial-time algorithm for learning decision trees that is optimally resilient to adversarial noise.
Our algorithm is furthermore proper, returning a hypothesis that is itself a decision tree.
arXiv Detail & Related papers (2021-05-08T04:54:12Z) - An Optimal Separation of Randomized and Quantum Query Complexity [67.19751155411075]
We prove that for every decision tree, the absolute values of the Fourier coefficients of a given order $ellsqrtbinomdell (1+log n)ell-1,$ sum to at most $cellsqrtbinomdell (1+log n)ell-1,$ where $n$ is the number of variables, $d$ is the tree depth, and $c>0$ is an absolute constant.
arXiv Detail & Related papers (2020-08-24T06:50:57Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z) - Provable guarantees for decision tree induction: the agnostic setting [16.784355746717562]
We give strengthened provable guarantees on the performance of widely employed and empirically successful sl top-down decision tree learnings.
We show that for all monotone functions$f$ and parameters $sin mathN$, theses construct a decision tree of size $stildeO((log s)/varepsilon2)$ that achieves error.
We complement our algorithmic guarantee with a near-matching $stildeOmega(log s)$ lower bound.
arXiv Detail & Related papers (2020-06-01T06:44:07Z) - Agnostic Q-learning with Function Approximation in Deterministic
Systems: Tight Bounds on Approximation Error and Sample Complexity [94.37110094442136]
We study the problem of agnostic $Q$-learning with function approximation in deterministic systems.
We show that if $delta = Oleft(rho/sqrtdim_Eright)$, then one can find the optimal policy using $Oleft(dim_Eright)$.
arXiv Detail & Related papers (2020-02-17T18:41:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.