NoisyCUR: An algorithm for two-cost budgeted matrix completion
- URL: http://arxiv.org/abs/2104.08026v1
- Date: Fri, 16 Apr 2021 10:39:56 GMT
- Title: NoisyCUR: An algorithm for two-cost budgeted matrix completion
- Authors: Dong Hu, Alex Gittens, and Malik Magdon-Ismail
- Abstract summary: Matrix completion is a ubiquitous tool in machine learning and data analysis.
We introduce a regression-based completion algorithm for this setting.
We experimentally verify the performance of our approach on both synthetic and real data sets.
- Score: 13.094997642327371
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Matrix completion is a ubiquitous tool in machine learning and data analysis.
Most work in this area has focused on the number of observations necessary to
obtain an accurate low-rank approximation. In practice, however, the cost of
observations is an important limiting factor, and experimentalists may have on
hand multiple modes of observation with differing noise-vs-cost trade-offs.
This paper considers matrix completion subject to such constraints: a budget is
imposed and the experimentalist's goal is to allocate this budget between two
sampling modalities in order to recover an accurate low-rank approximation.
Specifically, we consider that it is possible to obtain low noise, high cost
observations of individual entries or high noise, low cost observations of
entire columns. We introduce a regression-based completion algorithm for this
setting and experimentally verify the performance of our approach on both
synthetic and real data sets. When the budget is low, our algorithm outperforms
standard completion algorithms. When the budget is high, our algorithm has
comparable error to standard nuclear norm completion algorithms and requires
much less computational effort.
Related papers
- Perturb-and-Project: Differentially Private Similarities and Marginals [73.98880839337873]
We revisit the input perturbations framework for differential privacy where noise is added to the input $Ain mathcalS$.
We first design novel efficient algorithms to privately release pair-wise cosine similarities.
We derive a novel algorithm to compute $k$-way marginal queries over $n$ features.
arXiv Detail & Related papers (2024-06-07T12:07:16Z) - Robust Sparse Estimation for Gaussians with Optimal Error under Huber Contamination [42.526664955704746]
We study sparse estimation tasks in Huber's contamination model with a focus on mean estimation, PCA, and linear regression.
For each of these tasks, we give the first sample and computationally efficient robust estimators with optimal error guarantees.
At the technical level, we develop a novel multidimensional filtering method in the sparse regime that may find other applications.
arXiv Detail & Related papers (2024-03-15T15:51:27Z) - High-dimensional Contextual Bandit Problem without Sparsity [8.782204980889077]
We propose an explore-then-commit (EtC) algorithm to address this problem and examine its performance.
We derive the optimal rate of the ETC algorithm in terms of $T$ and show that this rate can be achieved by balancing exploration and exploitation.
We introduce an adaptive explore-then-commit (AEtC) algorithm that adaptively finds the optimal balance.
arXiv Detail & Related papers (2023-06-19T15:29:32Z) - Fast Optimal Locally Private Mean Estimation via Random Projections [58.603579803010796]
We study the problem of locally private mean estimation of high-dimensional vectors in the Euclidean ball.
We propose a new algorithmic framework, ProjUnit, for private mean estimation.
Our framework is deceptively simple: each randomizer projects its input to a random low-dimensional subspace, normalizes the result, and then runs an optimal algorithm.
arXiv Detail & Related papers (2023-06-07T14:07:35Z) - Improved Convergence of Score-Based Diffusion Models via Prediction-Correction [15.772322871598085]
Score-based generative models (SGMs) are powerful tools to sample from complex data distributions.
This paper addresses the issue by considering a version of the popular predictor-corrector scheme.
We first estimate the final distribution via an inexact Langevin dynamics and then revert the process.
arXiv Detail & Related papers (2023-05-23T15:29:09Z) - Provably Efficient Reinforcement Learning via Surprise Bound [66.15308700413814]
We propose a provably efficient reinforcement learning algorithm (both computationally and statistically) with general value function approximations.
Our algorithm achieves reasonable regret bounds when applied to both the linear setting and the sparse high-dimensional linear setting.
arXiv Detail & Related papers (2023-02-22T20:21:25Z) - Stochastic Direct Search Method for Blind Resource Allocation [6.574808513848414]
We study direct search (also known as pattern search) methods for linearly constrained and derivative-free optimization.
We show that direct search methods achieves finite regret in the deterministic and unconstrained case.
We propose a simple extension of direct search that achieves a regret upper-bound of the order of $T2/3$.
arXiv Detail & Related papers (2022-10-11T07:40:45Z) - Should All Proposals be Treated Equally in Object Detection? [110.27485090952385]
The complexity-precision trade-off of an object detector is a critical problem for resource constrained vision tasks.
It is hypothesized that improved detection efficiency requires a paradigm shift, towards the unequal processing of proposals.
This results in better utilization of available computational budget, enabling higher accuracy for the same FLOPS.
arXiv Detail & Related papers (2022-07-07T18:26:32Z) - Stochastic Hard Thresholding Algorithms for AUC Maximization [49.00683387735522]
We develop a hard thresholding algorithm for AUC in distributiond classification.
We conduct experiments to show the efficiency and effectiveness of the proposed algorithms.
arXiv Detail & Related papers (2020-11-04T16:49:29Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - Robust Mean Estimation in High Dimensions via $\ell_0$ Minimization [21.65637588606572]
We study the robust mean estimation problem in high dimensions, where $alpha 0.5$ fraction of the data points can be arbitrarily corrupted.
Motivated by compressive sensing, we formulate the robust mean estimation problem as the minimization of the $ell_p$ $(0p1)$.
Both synthetic and real data experiments demonstrate that the proposed algorithms significantly outperform state-of-the-art robust mean estimation methods.
arXiv Detail & Related papers (2020-08-21T00:19:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.