PAC Mode Estimation using PPR Martingale Confidence Sequences
- URL: http://arxiv.org/abs/2109.05047v1
- Date: Fri, 10 Sep 2021 18:11:38 GMT
- Title: PAC Mode Estimation using PPR Martingale Confidence Sequences
- Authors: Shubham Anand Jain, Sanit Gupta, Denil Mehta, Inderjeet Jayakumar
Nair, Rohan Shah, Jian Vora, Sushil Khyalia, Sourav Das, Vinay J. Ribeiro,
Shivaram Kalyanakrishnan
- Abstract summary: We consider the problem of correctly identifying the mode of a discrete distribution $mathcalP$ with sufficiently high probability.
We propose a generalisation to mode estimation, in which $mathcalP$ may take $K geq 2$ values.
Our resulting stopping rule, denoted PPR-ME, is optimal in its sample complexity up to a logarithmic factor.
- Score: 5.623190096715942
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of correctly identifying the mode of a discrete
distribution $\mathcal{P}$ with sufficiently high probability by observing a
sequence of i.i.d. samples drawn according to $\mathcal{P}$. This problem
reduces to the estimation of a single parameter when $\mathcal{P}$ has a
support set of size $K = 2$. Noting the efficiency of prior-posterior-ratio
(PPR) martingale confidence sequences for handling this special case, we
propose a generalisation to mode estimation, in which $\mathcal{P}$ may take $K
\geq 2$ values. We observe that the "one-versus-one" principle yields a more
efficient generalisation than the "one-versus-rest" alternative. Our resulting
stopping rule, denoted PPR-ME, is optimal in its sample complexity up to a
logarithmic factor. Moreover, PPR-ME empirically outperforms several other
competing approaches for mode estimation. We demonstrate the gains offered by
PPR-ME in two practical applications: (1) sample-based forecasting of the
winner in indirect election systems, and (2) efficient verification of smart
contracts in permissionless blockchains.
Related papers
- Computing Optimal Manipulations in Cryptographic Self-Selection Proof-of-Stake Protocols [3.3139913966251227]
We develop methods to provably nail $f(alpha,beta)$ for any desired $(alpha,beta)$ up to arbitrary precision.
One technical challenge is that natural sampling-based estimates of the mean of our target distribution are emphnot unbiased estimators.
arXiv Detail & Related papers (2024-06-21T16:20:39Z) - Sample-efficient Learning of Infinite-horizon Average-reward MDPs with General Function Approximation [53.17668583030862]
We study infinite-horizon average-reward Markov decision processes (AMDPs) in the context of general function approximation.
We propose a novel algorithmic framework named Local-fitted Optimization with OPtimism (LOOP)
We show that LOOP achieves a sublinear $tildemathcalO(mathrmpoly(d, mathrmsp(V*)) sqrtTbeta )$ regret, where $d$ and $beta$ correspond to AGEC and log-covering number of the hypothesis class respectively
arXiv Detail & Related papers (2024-04-19T06:24:22Z) - Efficiently Escaping Saddle Points for Non-Convex Policy Optimization [40.0986936439803]
Policy gradient (PG) is widely used in reinforcement learning due to its scalability and good performance.
We propose a variance-reduced second-order method that uses second-order information in the form of Hessian vector products (HVP) and converges to an approximate second-order stationary point (SOSP) with sample complexity of $tildeO(epsilon-3)$.
arXiv Detail & Related papers (2023-11-15T12:36:45Z) - Improved Convergence of Score-Based Diffusion Models via Prediction-Correction [15.772322871598085]
Score-based generative models (SGMs) are powerful tools to sample from complex data distributions.
This paper addresses the issue by considering a version of the popular predictor-corrector scheme.
We first estimate the final distribution via an inexact Langevin dynamics and then revert the process.
arXiv Detail & Related papers (2023-05-23T15:29:09Z) - Near Sample-Optimal Reduction-based Policy Learning for Average Reward
MDP [58.13930707612128]
This work considers the sample complexity of obtaining an $varepsilon$-optimal policy in an average reward Markov Decision Process (AMDP)
We prove an upper bound of $widetilde O(H varepsilon-3 ln frac1delta)$ samples per state-action pair, where $H := sp(h*)$ is the span of bias of any optimal policy, $varepsilon$ is the accuracy and $delta$ is the failure probability.
arXiv Detail & Related papers (2022-12-01T15:57:58Z) - Convergence for score-based generative modeling with polynomial
complexity [9.953088581242845]
We prove the first convergence guarantees for the core mechanic behind Score-based generative modeling.
Compared to previous works, we do not incur error that grows exponentially in time or that suffers from a curse of dimensionality.
We show that a predictor-corrector gives better convergence than using either portion alone.
arXiv Detail & Related papers (2022-06-13T14:57:35Z) - Under-Approximating Expected Total Rewards in POMDPs [68.8204255655161]
We consider the optimal expected total reward to reach a goal state in a partially observable Markov decision process (POMDP)
We use mixed-integer linear programming (MILP) to find such minimal probability shifts and experimentally show that our techniques scale quite well.
arXiv Detail & Related papers (2022-01-21T16:43:03Z) - Consistent Sufficient Explanations and Minimal Local Rules for
explaining regression and classification models [0.0]
We extend the notion of probabilistic Sufficient Explanations (P-SE)
The crux of P-SE is to compute the conditional probability of maintaining the same prediction.
We deal with non-binary features, without learning the distribution of $X$ nor having the model for making predictions.
arXiv Detail & Related papers (2021-11-08T17:27:52Z) - Permutation Compressors for Provably Faster Distributed Nonconvex
Optimization [68.8204255655161]
We show that the MARINA method of Gorbunov et al (2021) can be considered as a state-of-the-art method in terms of theoretical communication complexity.
Theory of MARINA to support the theory of potentially em correlated compressors, extends to the method beyond the classical independent compressors setting.
arXiv Detail & Related papers (2021-10-07T09:38:15Z) - Sharp Statistical Guarantees for Adversarially Robust Gaussian
Classification [54.22421582955454]
We provide the first result of the optimal minimax guarantees for the excess risk for adversarially robust classification.
Results are stated in terms of the Adversarial Signal-to-Noise Ratio (AdvSNR), which generalizes a similar notion for standard linear classification to the adversarial setting.
arXiv Detail & Related papers (2020-06-29T21:06:52Z) - Private Stochastic Non-Convex Optimization: Adaptive Algorithms and
Tighter Generalization Bounds [72.63031036770425]
We propose differentially private (DP) algorithms for bound non-dimensional optimization.
We demonstrate two popular deep learning methods on the empirical advantages over standard gradient methods.
arXiv Detail & Related papers (2020-06-24T06:01:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.