Agnostic Multi-Group Active Learning
- URL: http://arxiv.org/abs/2306.01922v1
- Date: Fri, 2 Jun 2023 21:24:13 GMT
- Title: Agnostic Multi-Group Active Learning
- Authors: Nick Rittler, Kamalika Chaudhuri
- Abstract summary: We consider a variant of this problem from the perspective of active learning, where the learner is endowed with the power to decide which examples are labeled from each distribution in the collection.
Our main challenge is that standard active learning techniques such as disagreement-based active learning do not directly apply to the multi-group learning objective.
We modify existing algorithms to provide a consistent active learning algorithm for an agnostic formulation of multi-group learning.
- Score: 24.97598179536084
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by the problem of improving classification accuracy on rare or hard
subsets of a population, there has been recent interest in models of learning
where the goal is to generalize to a collection of distributions, each
representing a ``group''. We consider a variant of this problem from the
perspective of active learning, where the learner is endowed with the power to
decide which examples are labeled from each distribution in the collection, and
the goal is to minimize the number of label queries while maintaining
PAC-learning guarantees. Our main challenge is that standard active learning
techniques such as disagreement-based active learning do not directly apply to
the multi-group learning objective. We modify existing algorithms to provide a
consistent active learning algorithm for an agnostic formulation of multi-group
learning, which given a collection of $G$ distributions and a hypothesis class
$\mathcal{H}$ with VC-dimension $d$, outputs an $\epsilon$-optimal hypothesis
using $\tilde{O}\left( (\nu^2/\epsilon^2+1) G d \theta_{\mathcal{G}}^2
\log^2(1/\epsilon) + G\log(1/\epsilon)/\epsilon^2 \right)$ label queries, where
$\theta_{\mathcal{G}}$ is the worst-case disagreement coefficient over the
collection. Roughly speaking, this guarantee improves upon the label complexity
of standard multi-group learning in regimes where disagreement-based active
learning algorithms may be expected to succeed, and the number of groups is not
too large. We also consider the special case where each distribution in the
collection is individually realizable with respect to $\mathcal{H}$, and
demonstrate $\tilde{O}\left( G d \theta_{\mathcal{G}} \log(1/\epsilon) \right)$
label queries are sufficient for learning in this case. We further give an
approximation result for the full agnostic case inspired by the group
realizable strategy.
Related papers
- Understanding Aggregations of Proper Learners in Multiclass Classification [4.422219522591412]
Multiclass learnability is known to exhibit a properness barrier.
Recent advances in binary classification have demonstrated that this requirement can be satisfied using aggregations of proper learners.
We show that a single ERM requires $Omega left(fracd_G ln (1 / delta)epsilonright)$ samples.
arXiv Detail & Related papers (2024-10-30T07:12:02Z) - Achieving More with Less: A Tensor-Optimization-Powered Ensemble Method [53.170053108447455]
Ensemble learning is a method that leverages weak learners to produce a strong learner.
We design a smooth and convex objective function that leverages the concept of margin, making the strong learner more discriminative.
We then compare our algorithm with random forests of ten times the size and other classical methods across numerous datasets.
arXiv Detail & Related papers (2024-08-06T03:42:38Z) - Fast Rates for Bandit PAC Multiclass Classification [73.17969992976501]
We study multiclass PAC learning with bandit feedback, where inputs are classified into one of $K$ possible labels and feedback is limited to whether or not the predicted labels are correct.
Our main contribution is in designing a novel learning algorithm for the agnostic $(varepsilon,delta)$PAC version of the problem.
arXiv Detail & Related papers (2024-06-18T08:54:04Z) - Agnostically Learning Multi-index Models with Queries [54.290489524576756]
We study the power of query access for the task of agnostic learning under the Gaussian distribution.
We show that query access gives significant runtime improvements over random examples for agnostically learning MIMs.
arXiv Detail & Related papers (2023-12-27T15:50:47Z) - Testable Learning with Distribution Shift [9.036777309376697]
We define a new model called testable learning with distribution shift.
We obtain provably efficient algorithms for certifying the performance of a classifier on a test distribution.
We give several positive results for learning concept classes such as halfspaces, intersections of halfspaces, and decision trees.
arXiv Detail & Related papers (2023-11-25T23:57:45Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Multi-Instance Partial-Label Learning: Towards Exploiting Dual Inexact
Supervision [53.530957567507365]
In some real-world tasks, each training sample is associated with a candidate label set that contains one ground-truth label and some false positive labels.
In this paper, we formalize such problems as multi-instance partial-label learning (MIPL)
Existing multi-instance learning algorithms and partial-label learning algorithms are suboptimal for solving MIPL problems.
arXiv Detail & Related papers (2022-12-18T03:28:51Z) - When are Local Queries Useful for Robust Learning? [25.832511407411637]
We study learning models where the learner is given more power through the use of local queries.
We give the first distribution-free algorithms that perform robust empirical risk minimization.
We finish by giving robust learning algorithms for halfspaces on $0,1n$ and then obtaining robustness guarantees for halfspaces in $mathbbRn$ against precision-bounded adversaries.
arXiv Detail & Related papers (2022-10-12T11:04:22Z) - Multi-group Agnostic PAC Learnability [7.9649015115693444]
We study "multi-group agnostic PAC learnability"
We provide a characterization of the loss functions for which such a predictor is guaranteed to exist.
Our results unify and extend previous positive and negative results from the multi-group fairness literature.
arXiv Detail & Related papers (2021-05-20T18:43:36Z) - Algorithms and SQ Lower Bounds for PAC Learning One-Hidden-Layer ReLU
Networks [48.32532049640782]
We study the problem of learning one-hidden-layer ReLU networks with $k$ hidden units on $mathbbRd$ under Gaussian marginals.
For the case of positive coefficients, we give the first-time algorithm for this learning problem for $k$ up to $tildeOOmega(sqrtlog d)$.
arXiv Detail & Related papers (2020-06-22T17:53:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.