Regret Minimization with Noisy Observations
- URL: http://arxiv.org/abs/2207.09435v1
- Date: Tue, 19 Jul 2022 17:48:54 GMT
- Title: Regret Minimization with Noisy Observations
- Authors: Mohammad Mahdian, Jieming Mao and Kangning Wang
- Abstract summary: In a typical optimization problem, the task is to pick one of a number of options with the lowest cost or the highest value.
In this paper, we study such scenarios using a minimization regret model.
We show that the na"ive algorithm of picking the highest observed value has regret arbitrarily worse than the optimum.
- Score: 18.24975895643299
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a typical optimization problem, the task is to pick one of a number of
options with the lowest cost or the highest value. In practice, these
cost/value quantities often come through processes such as measurement or
machine learning, which are noisy, with quantifiable noise distributions. To
take these noise distributions into account, one approach is to assume a prior
for the values, use it to build a posterior, and then apply standard stochastic
optimization to pick a solution. However, in many practical applications, such
prior distributions may not be available. In this paper, we study such
scenarios using a regret minimization model.
In our model, the task is to pick the highest one out of $n$ values. The
values are unknown and chosen by an adversary, but can be observed through
noisy channels, where additive noises are stochastically drawn from known
distributions. The goal is to minimize the regret of our selection, defined as
the expected difference between the highest and the selected value on the
worst-case choices of values. We show that the na\"ive algorithm of picking the
highest observed value has regret arbitrarily worse than the optimum, even when
$n = 2$ and the noises are unbiased in expectation. On the other hand, we
propose an algorithm which gives a constant-approximation to the optimal regret
for any $n$. Our algorithm is conceptually simple, computationally efficient,
and requires only minimal knowledge of the noise distributions.
Related papers
- Optimal Recovery Meets Minimax Estimation [20.84190812813031]
The goal is to design a numerical algorithm to construct an approximation $hat f$ to $f$ in a prescribed norm.
This paper handles this issue and determines the noise-level-aware (NLA) minimax rates for Besov classes when error is measured in an $L_q$-norm.
arXiv Detail & Related papers (2025-02-24T21:37:54Z) - Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv Detail & Related papers (2024-02-14T07:52:00Z) - Some Constructions of Private, Efficient, and Optimal $K$-Norm and Elliptic Gaussian Noise [54.34628844260993]
Differentially private computation often begins with a bound on some $d$-dimensional statistic's sensitivity.
For pure differential privacy, the $K$-norm mechanism can improve on this approach using a norm tailored to the statistic's sensitivity space.
This paper solves both problems for the simple statistics of sum, count, and vote.
arXiv Detail & Related papers (2023-09-27T17:09:36Z) - Optimizing the Noise in Self-Supervised Learning: from Importance
Sampling to Noise-Contrastive Estimation [80.07065346699005]
It is widely assumed that the optimal noise distribution should be made equal to the data distribution, as in Generative Adversarial Networks (GANs)
We turn to Noise-Contrastive Estimation which grounds this self-supervised task as an estimation problem of an energy-based model of the data.
We soberly conclude that the optimal noise may be hard to sample from, and the gain in efficiency can be modest compared to choosing the noise distribution equal to the data's.
arXiv Detail & Related papers (2023-01-23T19:57:58Z) - Stochastic Direct Search Method for Blind Resource Allocation [6.574808513848414]
We study direct search (also known as pattern search) methods for linearly constrained and derivative-free optimization.
We show that direct search methods achieves finite regret in the deterministic and unconstrained case.
We propose a simple extension of direct search that achieves a regret upper-bound of the order of $T2/3$.
arXiv Detail & Related papers (2022-10-11T07:40:45Z) - Greedy versus Map-based Optimized Adaptive Algorithms for
random-telegraph-noise mitigation by spectator qubits [6.305016513788048]
In a scenario where data-storage qubits are kept in isolation as far as possible, noise mitigation can still be done using additional noise probes.
We construct a theoretical model assuming projective measurements on the qubits, and derive the performance of different measurement and control strategies.
We show, analytically and numerically, that MOAAAR outperforms the Greedy algorithm, especially in the regime of high noise sensitivity of SQ.
arXiv Detail & Related papers (2022-05-25T08:25:10Z) - The Optimal Noise in Noise-Contrastive Learning Is Not What You Think [80.07065346699005]
We show that deviating from this assumption can actually lead to better statistical estimators.
In particular, the optimal noise distribution is different from the data's and even from a different family.
arXiv Detail & Related papers (2022-03-02T13:59:20Z) - Suboptimal Performance of the Bayes Optimal Algorithm in Frequentist Best Arm Identification [5.176134438571082]
We consider the fixed-budget best arm identification problem with rewards following normal distributions.
In this problem, the forecaster is given $K$ arms (or treatments) and $T$ time steps.
The algorithm's performance is evaluated by simple regret, reflecting the quality of the estimated best arm.
arXiv Detail & Related papers (2022-02-10T17:50:26Z) - Multinomial Logit Contextual Bandits: Provable Optimality and
Practicality [15.533842336139063]
We consider a sequential assortment selection problem where the user choice is given by a multinomial logit (MNL) choice model whose parameters are unknown.
We propose upper confidence bound based algorithms for this MNL contextual bandit.
We show that a simple variant of the algorithm achieves the optimal regret for a broad class of important applications.
arXiv Detail & Related papers (2021-03-25T15:42:25Z) - Learning with User-Level Privacy [61.62978104304273]
We analyze algorithms to solve a range of learning tasks under user-level differential privacy constraints.
Rather than guaranteeing only the privacy of individual samples, user-level DP protects a user's entire contribution.
We derive an algorithm that privately answers a sequence of $K$ adaptively chosen queries with privacy cost proportional to $tau$, and apply it to solve the learning tasks we consider.
arXiv Detail & Related papers (2021-02-23T18:25:13Z) - Lenient Regret for Multi-Armed Bandits [72.56064196252498]
We consider the Multi-Armed Bandit (MAB) problem, where an agent sequentially chooses actions and observes rewards for the actions it took.
While the majority of algorithms try to minimize the regret, i.e., the cumulative difference between the reward of the best action and the agent's action, this criterion might lead to undesirable results.
We suggest a new, more lenient, regret criterion that ignores suboptimality gaps smaller than some $epsilon$.
arXiv Detail & Related papers (2020-08-10T08:30:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.