Efficient One Sided Kolmogorov Approximation
        - URL: http://arxiv.org/abs/2207.07916v1
 - Date: Thu, 14 Jul 2022 10:03:02 GMT
 - Title: Efficient One Sided Kolmogorov Approximation
 - Authors: Liat Cohen, Tal Grinshpoun, Gera Weiss
 - Abstract summary: The main application that we examine is estimation of the probability missing deadlines in series-parallel schedules.
Since exact computation of these probabilities is NP-hard, we propose to use the algorithms described in this paper to obtain an approximation.
 - Score: 7.657378889055477
 - License: http://creativecommons.org/licenses/by/4.0/
 - Abstract:   We present an efficient algorithm that, given a discrete random variable $X$
and a number $m$, computes a random variable whose support is of size at most
$m$ and whose Kolmogorov distance from $X$ is minimal, also for the one-sided
Kolmogorov approximation. We present some variants of the algorithm, analyse
their correctness and computational complexity, and present a detailed
empirical evaluation that shows how they performs in practice. The main
application that we examine, which is our motivation for this work, is
estimation of the probability missing deadlines in series-parallel schedules.
Since exact computation of these probabilities is NP-hard, we propose to use
the algorithms described in this paper to obtain an approximation.
 
       
      
        Related papers
        - A Randomized Algorithm for Sparse PCA based on the Basic SDP Relaxation [1.3044677039636754]
We introduce a randomized approximation algorithm for SPCA, which is based on the basic SDP relaxation.<n>Our algorithm has an approximation ratio of at most the sparsity constant with high probability, if called enough times.<n>We demonstrate the efficacy of our algorithm through numerical results on real-world datasets.
arXiv  Detail & Related papers  (2025-07-12T05:43:56Z) - Pathwise optimization for bridge-type estimators and its applications [49.1574468325115]
Pathwise methods allow to efficiently compute the full path for penalized estimators.
We apply these algorithms to the penalized estimation of processes observed at discrete times.
arXiv  Detail & Related papers  (2024-12-05T10:38:29Z) - Perturb-and-Project: Differentially Private Similarities and Marginals [73.98880839337873]
We revisit the input perturbations framework for differential privacy where noise is added to the input $Ain mathcalS$.
We first design novel efficient algorithms to privately release pair-wise cosine similarities.
We derive a novel algorithm to compute $k$-way marginal queries over $n$ features.
arXiv  Detail & Related papers  (2024-06-07T12:07:16Z) - Robust Sparse Estimation for Gaussians with Optimal Error under Huber   Contamination [42.526664955704746]
We study sparse estimation tasks in Huber's contamination model with a focus on mean estimation, PCA, and linear regression.
For each of these tasks, we give the first sample and computationally efficient robust estimators with optimal error guarantees.
At the technical level, we develop a novel multidimensional filtering method in the sparse regime that may find other applications.
arXiv  Detail & Related papers  (2024-03-15T15:51:27Z) - Learning distributed representations with efficient SoftMax   normalization [3.8673630752805437]
We propose a linear-time approximation to compute the normalization constants of $rm SoftMax(XYT)$ for embedding vectors with bounded norms.<n>We show on some pre-trained embedding datasets that the proposed estimation method achieves higher or comparable accuracy with competing methods.<n>The proposed algorithm is interpretable and easily adapted to arbitrary embedding problems.
arXiv  Detail & Related papers  (2023-03-30T15:48:26Z) - Differentially-Private Hierarchical Clustering with Provable
  Approximation Guarantees [79.59010418610625]
We study differentially private approximation algorithms for hierarchical clustering.
We show strong lower bounds for the problem: that any $epsilon$-DP algorithm must exhibit $O(|V|2/ epsilon)$-additive error for an input dataset.
We propose a private $1+o(1)$ approximation algorithm which also recovers the blocks exactly.
arXiv  Detail & Related papers  (2023-01-31T19:14:30Z) - Private estimation algorithms for stochastic block models and mixture
  models [63.07482515700984]
General tools for designing efficient private estimation algorithms.
First efficient $(epsilon, delta)$-differentially private algorithm for both weak recovery and exact recovery.
arXiv  Detail & Related papers  (2023-01-11T09:12:28Z) - Privately Estimating a Gaussian: Efficient, Robust and Optimal [6.901744415870126]
We give efficient algorithms for privately estimating a Gaussian distribution in both pure and approximate differential privacy (DP) models.
In the pure DP setting, we give an efficient algorithm that estimates an unknown $d$-dimensional Gaussian distribution up to an arbitrary tiny total variation error.
For the special case of mean estimation, our algorithm achieves the optimal sample complexity of $widetilde O(d)$, improving on a $widetilde O(d1.5)$ bound from prior work.
arXiv  Detail & Related papers  (2022-12-15T18:27:39Z) - On the Sample Complexity of Representation Learning in Multi-task
  Bandits with Global and Local structure [77.60508571062958]
We investigate the sample complexity of learning the optimal arm for multi-task bandit problems.
Arms consist of two components: one that is shared across tasks (that we call representation) and one that is task-specific (that we call predictor)
We devise an algorithm OSRL-SC whose sample complexity approaches the lower bound, and scales at most as $H(Glog(delta_G)+ Xlog(delta_H))$, with $X,G,H$ being, respectively, the number of tasks, representations and predictors.
arXiv  Detail & Related papers  (2022-11-28T08:40:12Z) - Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time
  Guarantees [56.848265937921354]
Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy.
Many algorithms for IRL have an inherently nested structure.
We develop a novel single-loop algorithm for IRL that does not compromise reward estimation accuracy.
arXiv  Detail & Related papers  (2022-10-04T17:13:45Z) - Scalable Differentially Private Clustering via Hierarchically Separated
  Trees [82.69664595378869]
We show that our method computes a solution with cost at most $O(d3/2log n)cdot OPT + O(k d2 log2 n / epsilon2)$, where $epsilon$ is the privacy guarantee.
Although the worst-case guarantee is worse than that of state of the art private clustering methods, the algorithm we propose is practical.
arXiv  Detail & Related papers  (2022-06-17T09:24:41Z) - Robust Sparse Mean Estimation via Sum of Squares [42.526664955704746]
We study the problem of high-dimensional sparse mean estimation in the presence of an $epsilon$-fraction of adversarial outliers.
Our algorithms follow the Sum-of-Squares based, to algorithms approach.
arXiv  Detail & Related papers  (2022-06-07T16:49:54Z) - Grouped Variable Selection with Discrete Optimization: Computational and
  Statistical Perspectives [9.593208961737572]
We present a new algorithmic framework for grouped variable selection that is based on discrete mathematical optimization.
Our methodology covers both high-dimensional linear regression and non- additive modeling with sparse programming.
Our exact algorithm is based on a standalone branch-and-bound (BnB) framework, which can solve the associated mixed integer (MIP) problem to certified optimality.
arXiv  Detail & Related papers  (2021-04-14T19:21:59Z) - List-Decodable Mean Estimation in Nearly-PCA Time [50.79691056481693]
We study the fundamental task of list-decodable mean estimation in high dimensions.
Our algorithm runs in time $widetildeO(ndk)$ for all $k = O(sqrtd) cup Omega(d)$, where $n$ is the size of the dataset.
A variant of our algorithm has runtime $widetildeO(ndk)$ for all $k$, at the expense of an $O(sqrtlog k)$ factor in the recovery guarantee 
arXiv  Detail & Related papers  (2020-11-19T17:21:37Z) - Learning Sparse Classifiers: Continuous and Mixed Integer Optimization
  Perspectives [10.291482850329892]
Mixed integer programming (MIP) can be used to solve (to optimality) $ell_0$-regularized regression problems.
We propose two classes of scalable algorithms: an exact algorithm that can handlepapprox 50,000$ features in a few minutes, and approximate algorithms that can address instances with $papprox6$.
In addition, we present new estimation error bounds for $ell$-regularizeds.
arXiv  Detail & Related papers  (2020-01-17T18:47:02Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.