A Polynomial Time Algorithm for Learning Halfspaces with Tsybakov Noise
- URL: http://arxiv.org/abs/2010.01705v1
- Date: Sun, 4 Oct 2020 22:19:06 GMT
- Title: A Polynomial Time Algorithm for Learning Halfspaces with Tsybakov Noise
- Authors: Ilias Diakonikolas, Daniel M. Kane, Vasilis Kontonis, Christos Tzamos,
Nikos Zarifis
- Abstract summary: We study the problem of PAC learning homogeneous halfspaces in the presence of Tsybakov noise.
Our algorithm learns the true halfspace within any accuracy $epsilon$.
- Score: 55.45544638410858
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of PAC learning homogeneous halfspaces in the presence
of Tsybakov noise. In the Tsybakov noise model, the label of every sample is
independently flipped with an adversarially controlled probability that can be
arbitrarily close to $1/2$ for a fraction of the samples. {\em We give the
first polynomial-time algorithm for this fundamental learning problem.} Our
algorithm learns the true halfspace within any desired accuracy $\epsilon$ and
succeeds under a broad family of well-behaved distributions including
log-concave distributions. Prior to our work, the only previous algorithm for
this problem required quasi-polynomial runtime in $1/\epsilon$.
Our algorithm employs a recently developed reduction \cite{DKTZ20b} from
learning to certifying the non-optimality of a candidate halfspace. This prior
work developed a quasi-polynomial time certificate algorithm based on
polynomial regression. {\em The main technical contribution of the current
paper is the first polynomial-time certificate algorithm.} Starting from a
non-trivial warm-start, our algorithm performs a novel "win-win" iterative
process which, at each step, either finds a valid certificate or improves the
angle between the current halfspace and the true one. Our warm-start algorithm
for isotropic log-concave distributions involves a number of analytic tools
that may be of broader interest. These include a new efficient method for
reweighting the distribution in order to recenter it and a novel
characterization of the spectrum of the degree-$2$ Chow parameters.
Related papers
- Semi-Bandit Learning for Monotone Stochastic Optimization [20.776114616154242]
We provide a generic online learning algorithm for a class of "monotone" problems.
Our framework applies to several fundamental problems in optimization such as prophet, Pandora's box knapsack, inequality matchings and submodular optimization.
arXiv Detail & Related papers (2023-12-24T07:46:37Z) - Contrastive Moments: Unsupervised Halfspace Learning in Polynomial Time [8.419603619167813]
We give a Gaussian-time algorithm for learning high-dimensional halfspaces with margins in $d$-dimensional space to within desired TV distance.
Notably, our algorithm does not need labels and establishes unique (and efficient) identifiability of the hidden halfspace.
We improve on this by providing polytime guarantees based on Total Variation (TV) distance, in place of existing moment-bound guarantees that can be super-polynomial.
arXiv Detail & Related papers (2023-11-02T17:51:10Z) - A Strongly Polynomial Algorithm for Approximate Forster Transforms and
its Application to Halfspace Learning [56.86097988183311]
The Forster transform is a method of regularizing a dataset by placing it in em radial isotropic position while maintaining some of its essential properties.
arXiv Detail & Related papers (2022-12-06T14:32:02Z) - Higher degree sum-of-squares relaxations robust against oblivious
outliers [14.58610686291355]
We consider estimation models of the form $Y=X*+N$, where $X*$ is some $m$-dimensional signal we wish to recover.
We introduce a family of algorithms that under mild assumptions recover the signal $X*$ in all estimation problems for which there exists a sum-of-squares algorithm.
arXiv Detail & Related papers (2022-11-14T13:09:12Z) - Agnostic Proper Learning of Halfspaces under Gaussian Marginals [56.01192577666607]
We study the problem of agnostically learning halfspaces under the Gaussian.
Our main result is the em first proper learning algorithm for this problem.
arXiv Detail & Related papers (2021-02-10T18:40:44Z) - Finite-Time Analysis for Double Q-learning [50.50058000948908]
We provide the first non-asymptotic, finite-time analysis for double Q-learning.
We show that both synchronous and asynchronous double Q-learning are guaranteed to converge to an $epsilon$-accurate neighborhood of the global optimum.
arXiv Detail & Related papers (2020-09-29T18:48:21Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - Learning Halfspaces with Tsybakov Noise [50.659479930171585]
We study the learnability of halfspaces in the presence of Tsybakov noise.
We give an algorithm that achieves misclassification error $epsilon$ with respect to the true halfspace.
arXiv Detail & Related papers (2020-06-11T14:25:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.