Agnostically Learning Single-Index Models using Omnipredictors
- URL: http://arxiv.org/abs/2306.10615v1
- Date: Sun, 18 Jun 2023 18:40:07 GMT
- Title: Agnostically Learning Single-Index Models using Omnipredictors
- Authors: Aravind Gollakota and Parikshit Gopalan and Adam R. Klivans and
Konstantinos Stavropoulos
- Abstract summary: We give the first result for agnostically learning Single-Index Models (SIMs) with arbitrary monotone and Lipschitz activations.
We also provide new guarantees for standard algorithms like GLMtron and logistic regression in the agnostic setting.
- Score: 15.36798336447733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We give the first result for agnostically learning Single-Index Models (SIMs)
with arbitrary monotone and Lipschitz activations. All prior work either held
only in the realizable setting or required the activation to be known.
Moreover, we only require the marginal to have bounded second moments, whereas
all prior work required stronger distributional assumptions (such as
anticoncentration or boundedness). Our algorithm is based on recent work by
[GHK$^+$23] on omniprediction using predictors satisfying calibrated
multiaccuracy. Our analysis is simple and relies on the relationship between
Bregman divergences (or matching losses) and $\ell_p$ distances. We also
provide new guarantees for standard algorithms like GLMtron and logistic
regression in the agnostic setting.
Related papers
- Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Reinforcement Learning with Quasi-Hyperbolic Discounting [2.3999111269325266]
Quasi-Hyperbolic (QH) discounting is a simple alternative for modeling this bias.
Our work significantly advances the practical application of reinforcement learning.
arXiv Detail & Related papers (2024-09-16T06:00:42Z) - Robustly Learning Single-Index Models via Alignment Sharpness [40.886706402941435]
We study the problem of learning Single-Index Models under the $L2$ loss in the agnostic model.
We give an efficient learning algorithm, achieving a constant factor approximation to the optimal loss.
arXiv Detail & Related papers (2024-02-27T18:48:07Z) - Adaptive, Doubly Optimal No-Regret Learning in Strongly Monotone and Exp-Concave Games with Gradient Feedback [75.29048190099523]
Online gradient descent (OGD) is well known to be doubly optimal under strong convexity or monotonicity assumptions.
In this paper, we design a fully adaptive OGD algorithm, textsfAdaOGD, that does not require a priori knowledge of these parameters.
arXiv Detail & Related papers (2023-10-21T18:38:13Z) - Online Learning with Adversaries: A Differential-Inclusion Analysis [52.43460995467893]
We introduce an observation-matrix-based framework for fully asynchronous online Federated Learning with adversaries.
Our main result is that the proposed algorithm almost surely converges to the desired mean $mu.$
We derive this convergence using a novel differential-inclusion-based two-timescale analysis.
arXiv Detail & Related papers (2023-04-04T04:32:29Z) - Bayesian decision-making under misspecified priors with applications to
meta-learning [64.38020203019013]
Thompson sampling and other sequential decision-making algorithms are popular approaches to tackle explore/exploit trade-offs in contextual bandits.
We show that performance degrades gracefully with misspecified priors.
arXiv Detail & Related papers (2021-07-03T23:17:26Z) - Navigating to the Best Policy in Markov Decision Processes [68.8204255655161]
We investigate the active pure exploration problem in Markov Decision Processes.
Agent sequentially selects actions and, from the resulting system trajectory, aims at the best as fast as possible.
arXiv Detail & Related papers (2021-06-05T09:16:28Z) - Statistical-Query Lower Bounds via Functional Gradients [19.5924910463796]
We show that any statistical-query algorithm with tolerance $n- (1/epsilon)b$ must use at least $2nc epsilon$ queries for some constant $b.
Our results rule out general (as opposed to correlational) SQ learning algorithms, which is unusual for real-valued learning problems.
arXiv Detail & Related papers (2020-06-29T05:15:32Z) - Breaking the Sample Size Barrier in Model-Based Reinforcement Learning
with a Generative Model [50.38446482252857]
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator)
We first consider $gamma$-discounted infinite-horizon Markov decision processes (MDPs) with state space $mathcalS$ and action space $mathcalA$.
We prove that a plain model-based planning algorithm suffices to achieve minimax-optimal sample complexity given any target accuracy level.
arXiv Detail & Related papers (2020-05-26T17:53:18Z) - A General Theory of the Stochastic Linear Bandit and Its Applications [8.071506311915398]
We introduce a general analysis framework and a family of algorithms for the linear bandit problem.
Our new notion of optimism in expectation gives rise to a new algorithm, called sieved greedy (SG) that reduces the overexploration problem in OFUL.
In addition to proving that SG is theoretically rate optimal, our empirical simulations show that SG outperforms existing benchmarks such as greedy, OFUL, and TS.
arXiv Detail & Related papers (2020-02-12T18:54:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.