Causal Matrix Completion
- URL: http://arxiv.org/abs/2109.15154v1
- Date: Thu, 30 Sep 2021 14:17:56 GMT
- Title: Causal Matrix Completion
- Authors: Anish Agarwal, Munther Dahleh, Devavrat Shah, Dennis Shen
- Abstract summary: Matrix completion is the study of recovering an underlying matrix from a sparse subset of noisy observations.
Traditionally, it is assumed that the entries of the matrix are "missing completely at random"
- Score: 15.599296461516984
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Matrix completion is the study of recovering an underlying matrix from a
sparse subset of noisy observations. Traditionally, it is assumed that the
entries of the matrix are "missing completely at random" (MCAR), i.e., each
entry is revealed at random, independent of everything else, with uniform
probability. This is likely unrealistic due to the presence of "latent
confounders", i.e., unobserved factors that determine both the entries of the
underlying matrix and the missingness pattern in the observed matrix. For
example, in the context of movie recommender systems -- a canonical application
for matrix completion -- a user who vehemently dislikes horror films is
unlikely to ever watch horror films. In general, these confounders yield
"missing not at random" (MNAR) data, which can severely impact any inference
procedure that does not correct for this bias. We develop a formal causal model
for matrix completion through the language of potential outcomes, and provide
novel identification arguments for a variety of causal estimands of interest.
We design a procedure, which we call "synthetic nearest neighbors" (SNN), to
estimate these causal estimands. We prove finite-sample consistency and
asymptotic normality of our estimator. Our analysis also leads to new
theoretical results for the matrix completion literature. In particular, we
establish entry-wise, i.e., max-norm, finite-sample consistency and asymptotic
normality results for matrix completion with MNAR data. As a special case, this
also provides entry-wise bounds for matrix completion with MCAR data. Across
simulated and real data, we demonstrate the efficacy of our proposed estimator.
Related papers
- Negative Binomial Matrix Completion [5.5415918072761805]
Matrix completion focuses on recovering missing or incomplete information in matrices.
We introduce NB matrix completion by proposing a nuclear-norm regularized model that can be solved by proximal descent gradient.
In our experiments, we demonstrate that the NB model outperforms Poisson matrix completion in various noise and missing data settings on real data.
arXiv Detail & Related papers (2024-08-28T19:43:48Z) - Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Concentration properties of fractional posterior in 1-bit matrix completion [0.0]
This work specifically addresses the scenario of binary observations, often termed as 1-bit matrix completion.
We tackle this gap by considering a general, non-uniform sampling scheme and providing theoretical assurances on the efficacy of the fractional posterior.
Our results are comparable to those found in the frequentist literature, yet demand fewer restrictive assumptions.
arXiv Detail & Related papers (2024-04-13T11:22:53Z) - Statistical Inference For Noisy Matrix Completion Incorporating Auxiliary Information [3.9748528039819977]
This paper investigates statistical inference for noisy matrix completion in a semi-supervised model.
We apply an iterative least squares (LS) estimation approach in our considered context.
We show that our method only needs a few iterations, and the resulting entry-wise estimators of the low-rank matrix and the coefficient matrix are guaranteed to have normal distributions.
arXiv Detail & Related papers (2024-03-22T01:06:36Z) - The Ordered Matrix Dirichlet for Modeling Ordinal Dynamics [54.96229007229786]
We propose the Ordered Matrix Dirichlet (OMD) to map latent states to observed action types.
Models built on the OMD recover interpretable latent states and show superior forecasting performance in few-shot settings.
arXiv Detail & Related papers (2022-12-08T08:04:26Z) - Test Set Sizing Via Random Matrix Theory [91.3755431537592]
This paper uses techniques from Random Matrix Theory to find the ideal training-testing data split for a simple linear regression.
It defines "ideal" as satisfying the integrity metric, i.e. the empirical model error is the actual measurement noise.
This paper is the first to solve for the training and test size for any model in a way that is truly optimal.
arXiv Detail & Related papers (2021-12-11T13:18:33Z) - Leveraged Matrix Completion with Noise [84.20092979053119]
We show that we can provably recover an unknown $ntimes n$ matrix of rank $r$ from just about $mathcalO(nrlog2 (n))$ entries.
Our proofs are supported by a novel approach that phrases sufficient optimality conditions based on the Golfing Scheme.
arXiv Detail & Related papers (2020-11-11T16:25:45Z) - Adversarial Robust Low Rank Matrix Estimation: Compressed Sensing and Matrix Completion [2.0257616108612373]
We deal with matrix compressed sensing, including lasso as a partial problem, and matrix completion.
We propose a simple unified approach based on a combination of the Huber loss function and the nuclear norm penalization.
arXiv Detail & Related papers (2020-10-25T02:32:07Z) - Median Matrix Completion: from Embarrassment to Optimality [16.667260586938234]
We consider matrix completion with absolute deviation loss and obtain an estimator of the median matrix.
Despite several appealing properties of median, the non-smooth absolute deviation loss leads to computational challenge.
We propose a novel refinement step, which turns such inefficient estimators into a rate (near-optimal) matrix completion procedure.
arXiv Detail & Related papers (2020-06-18T10:01:22Z) - Robust Compressed Sensing using Generative Models [98.64228459705859]
In this paper we propose an algorithm inspired by the Median-of-Means (MOM)
Our algorithm guarantees recovery for heavy-tailed data, even in the presence of outliers.
arXiv Detail & Related papers (2020-06-16T19:07:41Z) - Relative Error Bound Analysis for Nuclear Norm Regularized Matrix Completion [101.83262280224729]
We develop a relative error bound for nuclear norm regularized matrix completion.
We derive a relative upper bound for recovering the best low-rank approximation of the unknown matrix.
arXiv Detail & Related papers (2015-04-26T13:12:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.