On the Comparison between Cyclic Sampling and Random Reshuffling
- URL: http://arxiv.org/abs/2104.12112v1
- Date: Sun, 25 Apr 2021 09:29:43 GMT
- Title: On the Comparison between Cyclic Sampling and Random Reshuffling
- Authors: Xinmeng Huang, Kun Yuan, Xianghui Mao, Wotao Yin
- Abstract summary: Cyclic sampling draws the samples in a fixed, cyclic order, which is less robust than reshuffling the samples periodically.
Existing works have established worst case convergence rates for cyclic sampling, which are generally worse than that of random reshuffling.
In this paper, we found a certain cyclic order can be much faster than reshuffling and one can discover it at a low cost.
- Score: 27.27774034428748
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When applying a stochastic/incremental algorithm, one must choose the order
to draw samples. Among the most popular approaches are cyclic sampling and
random reshuffling, which are empirically faster and more cache-friendly than
uniform-iid-sampling. Cyclic sampling draws the samples in a fixed, cyclic
order, which is less robust than reshuffling the samples periodically. Indeed,
existing works have established worst case convergence rates for cyclic
sampling, which are generally worse than that of random reshuffling. In this
paper, however, we found a certain cyclic order can be much faster than
reshuffling and one can discover it at a low cost!
Studying and comparing different sampling orders typically require new
analytic techniques. In this paper, we introduce a norm, which is defined based
on the sampling order, to measure the distance to solution. Applying this
technique on proximal Finito/MISO algorithm allows us to identify the optimal
fixed ordering, which can beat random reshuffling by a factor up to log(n)/n in
terms of the best-known upper bounds. We also propose a strategy to discover
the optimal fixed ordering numerically. The established rates are
state-of-the-art compared to previous works.
Related papers
- Sampling in CMA-ES: Low Numbers of Low Discrepancy Points [0.0]
We show that iterating through small, fixed sets of low-discrepancy points can still perform better than the default uniform distribution.
For lower dimensionalities, we find that using as little as 32 unique discrepancy points performs similar or better than uniform sampling.
arXiv Detail & Related papers (2024-09-24T10:04:55Z) - Stochastic optimization with arbitrary recurrent data sampling [2.1485350418225244]
Most commonly used data sampling algorithms are under mild assumptions.
We show that for a particular class of recurrent optimization algorithms, we do not need any other property.
We show that convergence can be accelerated by selecting sampling algorithms that cover the data set.
arXiv Detail & Related papers (2024-01-15T14:04:50Z) - Stable generative modeling using Schrödinger bridges [0.22499166814992438]
We propose a generative model combining Schr"odinger bridges and Langevin dynamics.
Our framework can be naturally extended to generate conditional samples and to Bayesian inference problems.
arXiv Detail & Related papers (2024-01-09T06:15:45Z) - Sampling without Replacement Leads to Faster Rates in Finite-Sum Minimax
Optimization [12.794526174053134]
We analyze convergence rates of gradient algorithms for smooth finite-sum minimax optimization.
We show that, for many such algorithms, sampling points without replacement leads to faster convergence compared to sampling with replacement.
arXiv Detail & Related papers (2022-06-07T00:37:37Z) - Minibatch vs Local SGD with Shuffling: Tight Convergence Bounds and
Beyond [63.59034509960994]
We study shuffling-based variants: minibatch and local Random Reshuffling, which draw gradients without replacement.
For smooth functions satisfying the Polyak-Lojasiewicz condition, we obtain convergence bounds which show that these shuffling-based variants converge faster than their with-replacement counterparts.
We propose an algorithmic modification called synchronized shuffling that leads to convergence rates faster than our lower bounds in near-homogeneous settings.
arXiv Detail & Related papers (2021-10-20T02:25:25Z) - Adaptive Sampling for Heterogeneous Rank Aggregation from Noisy Pairwise
Comparisons [85.5955376526419]
In rank aggregation problems, users exhibit various accuracy levels when comparing pairs of items.
We propose an elimination-based active sampling strategy, which estimates the ranking of items via noisy pairwise comparisons.
We prove that our algorithm can return the true ranking of items with high probability.
arXiv Detail & Related papers (2021-10-08T13:51:55Z) - Towards Sample-Optimal Compressive Phase Retrieval with Sparse and
Generative Priors [59.33977545294148]
We show that $O(k log L)$ samples suffice to guarantee that the signal is close to any vector that minimizes an amplitude-based empirical loss function.
We adapt this result to sparse phase retrieval, and show that $O(s log n)$ samples are sufficient for a similar guarantee when the underlying signal is $s$-sparse and $n$-dimensional.
arXiv Detail & Related papers (2021-06-29T12:49:54Z) - Local policy search with Bayesian optimization [73.0364959221845]
Reinforcement learning aims to find an optimal policy by interaction with an environment.
Policy gradients for local search are often obtained from random perturbations.
We develop an algorithm utilizing a probabilistic model of the objective function and its gradient.
arXiv Detail & Related papers (2021-06-22T16:07:02Z) - Non-Adaptive Adaptive Sampling on Turnstile Streams [57.619901304728366]
We give the first relative-error algorithms for column subset selection, subspace approximation, projective clustering, and volume on turnstile streams that use space sublinear in $n$.
Our adaptive sampling procedure has a number of applications to various data summarization problems that either improve state-of-the-art or have only been previously studied in the more relaxed row-arrival model.
arXiv Detail & Related papers (2020-04-23T05:00:21Z) - The Simulator: Understanding Adaptive Sampling in the
Moderate-Confidence Regime [52.38455827779212]
We propose a novel technique for analyzing adaptive sampling called the em Simulator.
We prove the first instance-based lower bounds the top-k problem which incorporate the appropriate log-factors.
Our new analysis inspires a simple and near-optimal for the best-arm and top-k identification, the first em practical of its kind for the latter problem.
arXiv Detail & Related papers (2017-02-16T23:42:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.