Learning Halfspaces with Tsybakov Noise
- URL: http://arxiv.org/abs/2006.06467v1
- Date: Thu, 11 Jun 2020 14:25:02 GMT
- Title: Learning Halfspaces with Tsybakov Noise
- Authors: Ilias Diakonikolas, Vasilis Kontonis, Christos Tzamos, Nikos Zarifis
- Abstract summary: We study the learnability of halfspaces in the presence of Tsybakov noise.
We give an algorithm that achieves misclassification error $epsilon$ with respect to the true halfspace.
- Score: 50.659479930171585
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the efficient PAC learnability of halfspaces in the presence of
Tsybakov noise. In the Tsybakov noise model, each label is independently
flipped with some probability which is controlled by an adversary. This noise
model significantly generalizes the Massart noise model, by allowing the
flipping probabilities to be arbitrarily close to $1/2$ for a fraction of the
samples. Our main result is the first non-trivial PAC learning algorithm for
this problem under a broad family of structured distributions -- satisfying
certain concentration and (anti-)anti-concentration properties -- including
log-concave distributions. Specifically, we given an algorithm that achieves
misclassification error $\epsilon$ with respect to the true halfspace, with
quasi-polynomial runtime dependence in $1/\epsilin$. The only previous upper
bound for this problem -- even for the special case of log-concave
distributions -- was doubly exponential in $1/\epsilon$ (and follows via the
naive reduction to agnostic learning). Our approach relies on a novel
computationally efficient procedure to certify whether a candidate solution is
near-optimal, based on semi-definite programming. We use this certificate
procedure as a black-box and turn it into an efficient learning algorithm by
searching over the space of halfspaces via online convex optimization.
Related papers
- Efficient Active Learning Halfspaces with Tsybakov Noise: A Non-convex Optimization Approach [28.518140532315776]
We study the problem of computationally label efficient active learning.
In this paper, we prove any approximate.
first-order stationary light point of a structured un-labeled data set.
arXiv Detail & Related papers (2023-10-23T23:55:28Z) - Near-Optimal Bounds for Learning Gaussian Halfspaces with Random
Classification Noise [50.64137465792738]
We show that any efficient SQ algorithm for the problem requires sample complexity at least $Omega(d1/2/(maxp, epsilon)2)$.
Our lower bound suggests that this quadratic dependence on $1/epsilon$ is inherent for efficient algorithms.
arXiv Detail & Related papers (2023-07-13T18:59:28Z) - Boosting in the Presence of Massart Noise [49.72128048499074]
We study the problem of boosting the accuracy of a weak learner in the (distribution-independent) PAC model with Massart noise.
Our main result is the first computationally efficient boosting algorithm in the presence of Massart noise.
As a simple application of our positive result, we give first efficient Massart learner for unions of high-dimensional rectangles.
arXiv Detail & Related papers (2021-06-14T22:21:25Z) - Improved Algorithms for Efficient Active Learning Halfspaces with
Massart and Tsybakov noise [29.890039126644776]
We develop a PAC active learning algorithm for $d$-dimensional homogeneous halfspaces that can tolerate Massart noisecitepmassart2006risk and Tsybakov noiseciteptsybakov2004.
Under the more challenging Tsybakov noise condition, we identify two subfamilies of noise conditions, under which our algorithm achieves computational efficiency and provide label complexity guarantees strictly lower than passive learning algorithms.
arXiv Detail & Related papers (2021-02-10T08:17:17Z) - A Polynomial Time Algorithm for Learning Halfspaces with Tsybakov Noise [55.45544638410858]
We study the problem of PAC learning homogeneous halfspaces in the presence of Tsybakov noise.
Our algorithm learns the true halfspace within any accuracy $epsilon$.
arXiv Detail & Related papers (2020-10-04T22:19:06Z) - The Complexity of Adversarially Robust Proper Learning of Halfspaces
with Agnostic Noise [67.27523616312428]
We study the computational complexity of adversarially robust proper learning of halfspaces in the distribution-independent PAC model.
We give a computationally efficient learning algorithm and a nearly matching computational hardness result for this problem.
arXiv Detail & Related papers (2020-07-30T04:18:51Z) - Attribute-Efficient Learning of Halfspaces with Malicious Noise:
Near-Optimal Label Complexity and Noise Tolerance [21.76197397540397]
This paper is concerned with computationally efficient learning of homogeneous sparse halfspaces in $mathbbRd$ under noise.
We show that the sample complexity is $tildeObig(frac 1 epsilon s2 log5 d big)$ which also enjoys the attribute efficiency.
arXiv Detail & Related papers (2020-06-06T04:57:39Z) - Efficiently Learning Adversarially Robust Halfspaces with Noise [69.01459748050453]
We study the problem of learning adversarially robust halfspaces in the distribution-independent setting.
In the realizable setting, we provide necessary and sufficient conditions on the adversarial perturbation sets under which halfspaces are efficiently robustly learnable.
arXiv Detail & Related papers (2020-05-15T17:13:54Z) - Learning Halfspaces with Massart Noise Under Structured Distributions [46.37328271312332]
We solve the problem of learningspaces with Massart noise in the distribution-specific PAC model.
We give the first efficient algorithm for a broad family problem with respect to log-con.
arXiv Detail & Related papers (2020-02-13T17:02:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.