Bayesian matrix completion with a spectral scaled Student prior:
theoretical guarantee and efficient sampling
- URL: http://arxiv.org/abs/2104.08191v1
- Date: Fri, 16 Apr 2021 16:03:43 GMT
- Title: Bayesian matrix completion with a spectral scaled Student prior:
theoretical guarantee and efficient sampling
- Authors: The Tien Mai
- Abstract summary: A spectral scaled Student prior is exploited to favour the underlying low-rank structure of the data matrix.
We show that our approach enjoys a minimax-optimal oracle inequality which guarantees that our method works well under model misspecification.
- Score: 0.30458514384586394
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the problem of matrix completion in this paper. A spectral scaled
Student prior is exploited to favour the underlying low-rank structure of the
data matrix. Importantly, we provide a thorough theoretical investigation for
our approach, while such an analysis is hard to obtain and limited in
theoretical understanding of Bayesian matrix completion. More precisely, we
show that our Bayesian approach enjoys a minimax-optimal oracle inequality
which guarantees that our method works well under model misspecification and
under general sampling distribution. Interestingly, we also provide efficient
gradient-based sampling implementations for our approach by using Langevin
Monte Carlo which is novel in Bayesian matrix completion. More specifically, we
show that our algorithms are significantly faster than Gibbs sampler in this
problem. To illustrate the attractive features of our inference strategy, some
numerical simulations are conducted and an application to image inpainting is
demonstrated.
Related papers
- Low-rank Bayesian matrix completion via geodesic Hamiltonian Monte Carlo on Stiefel manifolds [0.18416014644193066]
We present a new sampling-based approach for enabling efficient computation of low-rank Bayesian matrix completion.
We show that our approach resolves the sampling difficulties encountered by standard Gibbs samplers for the common two-matrix factorization used in matrix completion.
Numerical examples demonstrate superior sampling performance, including better mixing and faster convergence to a stationary distribution.
arXiv Detail & Related papers (2024-10-27T03:12:53Z) - On Discriminative Probabilistic Modeling for Self-Supervised Representation Learning [85.75164588939185]
We study the discriminative probabilistic modeling problem on a continuous domain for (multimodal) self-supervised representation learning.
We conduct generalization error analysis to reveal the limitation of current InfoNCE-based contrastive loss for self-supervised representation learning.
arXiv Detail & Related papers (2024-10-11T18:02:46Z) - Concentration properties of fractional posterior in 1-bit matrix completion [0.0]
This work specifically addresses the scenario of binary observations, often termed as 1-bit matrix completion.
We tackle this gap by considering a general, non-uniform sampling scheme and providing theoretical assurances on the efficacy of the fractional posterior.
Our results are comparable to those found in the frequentist literature, yet demand fewer restrictive assumptions.
arXiv Detail & Related papers (2024-04-13T11:22:53Z) - Regularization by denoising: Bayesian model and Langevin-within-split
Gibbs sampling [6.453497703172228]
This paper introduces a Bayesian framework for image inversion by deriving a probabilistic counterpart to the regularization-by-denoising (RED) paradigm.
It implements a Monte Carlo algorithm specifically tailored for sampling from the resulting posterior distribution, based on anally exact data augmentation (AXDA)
The proposed algorithm is an approximate instance of split Gibbs sampling (SGS) which embeds one Langevin Monte Carlo step.
arXiv Detail & Related papers (2024-02-19T17:12:16Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Spectral Entry-wise Matrix Estimation for Low-Rank Reinforcement
Learning [53.445068584013896]
We study matrix estimation problems arising in reinforcement learning (RL) with low-rank structure.
In low-rank bandits, the matrix to be recovered specifies the expected arm rewards, and for low-rank Markov Decision Processes (MDPs), it may for example characterize the transition kernel of the MDP.
We show that simple spectral-based matrix estimation approaches efficiently recover the singular subspaces of the matrix and exhibit nearly-minimal entry-wise error.
arXiv Detail & Related papers (2023-10-10T17:06:41Z) - Large-scale Bayesian Structure Learning for Gaussian Graphical Models using Marginal Pseudo-likelihood [0.26249027950824516]
We introduce two novel Markov chain Monte Carlo (MCMC) search algorithms with a significantly lower computational cost than leading Bayesian approaches.
These algorithms can deliver reliable results in mere minutes on standard computers, even for large-scale problems with one thousand variables.
We also illustrate the practical utility of our methods on medium and large-scale applications from human and mice gene expression studies.
arXiv Detail & Related papers (2023-06-30T20:37:40Z) - Matrix Completion from General Deterministic Sampling Patterns [28.116011361245224]
We establish theoretical guarantee for the exact and approximate low-rank matrix completion problems.
We show that the algorithm can be successful as the observation graph is well-connected and has similar node degrees.
arXiv Detail & Related papers (2023-06-04T07:01:31Z) - On the Benefits of Large Learning Rates for Kernel Methods [110.03020563291788]
We show that a phenomenon can be precisely characterized in the context of kernel methods.
We consider the minimization of a quadratic objective in a separable Hilbert space, and show that with early stopping, the choice of learning rate influences the spectral decomposition of the obtained solution.
arXiv Detail & Related papers (2022-02-28T13:01:04Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - Online Dense Subgraph Discovery via Blurred-Graph Feedback [87.9850024070244]
We introduce a novel learning problem for dense subgraph discovery.
We first propose a edge-time algorithm that obtains a nearly-optimal solution with high probability.
We then design a more scalable algorithm with a theoretical guarantee.
arXiv Detail & Related papers (2020-06-24T11:37:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.