Rejection sampling from shape-constrained distributions in sublinear
time
- URL: http://arxiv.org/abs/2105.14166v1
- Date: Sat, 29 May 2021 01:00:42 GMT
- Title: Rejection sampling from shape-constrained distributions in sublinear
time
- Authors: Sinho Chewi, Patrik Gerber, Chen Lu, Thibaut Le Gouic, Philippe
Rigollet
- Abstract summary: We study the query complexity of rejection sampling in a minimax framework for various classes of discrete distributions.
Our results provide new algorithms for sampling whose complexity scales sublinearly with the alphabet size.
- Score: 14.18847457501901
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the task of generating exact samples from a target distribution,
known up to normalization, over a finite alphabet. The classical algorithm for
this task is rejection sampling, and although it has been used in practice for
decades, there is surprisingly little study of its fundamental limitations. In
this work, we study the query complexity of rejection sampling in a minimax
framework for various classes of discrete distributions. Our results provide
new algorithms for sampling whose complexity scales sublinearly with the
alphabet size. When applied to adversarial bandits, we show that a slight
modification of the Exp3 algorithm reduces the per-iteration complexity from
$\mathcal O(K)$ to $\mathcal O(\log^2 K)$, where $K$ is the number of arms.
Related papers
- Optimal Multi-Distribution Learning [88.3008613028333]
Multi-distribution learning seeks to learn a shared model that minimizes the worst-case risk across $k$ distinct data distributions.
We propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon2.
arXiv Detail & Related papers (2023-12-08T16:06:29Z) - Information-Computation Tradeoffs for Learning Margin Halfspaces with
Random Classification Noise [50.64137465792738]
We study the problem of PAC $gamma$-margin halfspaces with Random Classification Noise.
We establish an information-computation tradeoff suggesting an inherent gap between the sample complexity of the problem and the sample complexity of computationally efficient algorithms.
arXiv Detail & Related papers (2023-06-28T16:33:39Z) - On-Demand Sampling: Learning Optimally from Multiple Distributions [63.20009081099896]
Social and real-world considerations have given rise to multi-distribution learning paradigms.
We establish the optimal sample complexity of these learning paradigms and give algorithms that meet this sample complexity.
Our algorithm design and analysis are enabled by our extensions of online learning techniques for solving zero-sum games.
arXiv Detail & Related papers (2022-10-22T19:07:26Z) - Undersampling is a Minimax Optimal Robustness Intervention in
Nonparametric Classification [28.128464387420216]
We show that learning is fundamentally constrained by a lack of minority group samples.
In particular, in the case of label shift we show that there is always an undersampling algorithm that is minimax optimal.
arXiv Detail & Related papers (2022-05-26T00:35:11Z) - A Proximal Algorithm for Sampling from Non-convex Potentials [14.909442791255042]
We consider problems with non-smooth potentials that lack smoothness.
Rather than smooth, the potentials are assumed to be semi-smooth or multiple multiplesmooth functions.
We develop a special case Gibbs sampling known as the alternating sampling framework.
arXiv Detail & Related papers (2022-05-20T13:58:46Z) - Lattice-Based Methods Surpass Sum-of-Squares in Clustering [98.46302040220395]
Clustering is a fundamental primitive in unsupervised learning.
Recent work has established lower bounds against the class of low-degree methods.
We show that, perhaps surprisingly, this particular clustering model textitdoes not exhibit a statistical-to-computational gap.
arXiv Detail & Related papers (2021-12-07T18:50:17Z) - Improved Algorithms for Agnostic Pool-based Active Classification [20.12178157010804]
We consider active learning for binary classification in the agnostic pool-based setting.
Our algorithm is superior to state of the art active learning algorithms on image classification datasets.
arXiv Detail & Related papers (2021-05-13T18:24:30Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - An Empirical Process Approach to the Union Bound: Practical Algorithms
for Combinatorial and Linear Bandits [34.06611065493047]
This paper proposes near-optimal algorithms for the pure-exploration linear bandit problem in the fixed confidence and fixed budget settings.
We provide an algorithm whose sample complexity scales with the geometry of the instance and avoids an explicit union bound over the number of arms.
We also propose the first algorithm for linear bandits in the the fixed budget setting.
arXiv Detail & Related papers (2020-06-21T00:56:33Z) - Nearly Linear Row Sampling Algorithm for Quantile Regression [54.75919082407094]
We give a row sampling algorithm for the quantile loss function with sample complexity nearly linear in the dimensionality of the data.
Based upon our row sampling algorithm, we give the fastest known algorithm for quantile regression and a graph sparsification algorithm for balanced directed graphs.
arXiv Detail & Related papers (2020-06-15T13:40:07Z) - Learning Structured Distributions From Untrusted Batches: Faster and
Simpler [26.57313255285721]
We revisit the problem of learning from untrusted batches introduced by Qiao and Valiant [QV17].
We find an appealing way to synthesize the techniques of [JO19] and [CLM19] to give the best of both worlds: an algorithm which runs in monotone time and can exploit structure in the underlying distribution to achieve sublinear sample.
arXiv Detail & Related papers (2020-02-24T18:32:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.