Probabilistic semi-nonnegative matrix factorization: a Skellam-based
framework
- URL: http://arxiv.org/abs/2107.03317v1
- Date: Wed, 7 Jul 2021 15:56:22 GMT
- Title: Probabilistic semi-nonnegative matrix factorization: a Skellam-based
framework
- Authors: Benoit Fuentes, Ga\"el Richard
- Abstract summary: We present a new probabilistic model to address semi-nonnegative matrix factorization (SNMF), called Skellam-SNMF.
It is a hierarchical generative model consisting of prior components, Skellam-distributed hidden variables and observed data.
Two inference algorithms are derived: Expectation-Maximization (EM) algorithm for maximum empha posteriori estimation and Vari Bayes EM (VBEM) for full Bayesian inference.
- Score: 0.7310043452300736
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We present a new probabilistic model to address semi-nonnegative matrix
factorization (SNMF), called Skellam-SNMF. It is a hierarchical generative
model consisting of prior components, Skellam-distributed hidden variables and
observed data. Two inference algorithms are derived: Expectation-Maximization
(EM) algorithm for maximum \emph{a posteriori} estimation and Variational Bayes
EM (VBEM) for full Bayesian inference, including the estimation of parameters
prior distribution. From this Skellam-based model, we also introduce a new
divergence $\mathcal{D}$ between a real-valued target data $x$ and two
nonnegative parameters $\lambda_{0}$ and $\lambda_{1}$ such that
$\mathcal{D}\left(x\mid\lambda_{0},\lambda_{1}\right)=0\Leftrightarrow
x=\lambda_{0}-\lambda_{1}$, which is a generalization of the Kullback-Leibler
(KL) divergence. Finally, we conduct experimental studies on those new
algorithms in order to understand their behavior and prove that they can
outperform the classic SNMF approach on real data in a task of automatic
clustering.
Related papers
- Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Dynamical System Identification, Model Selection and Model Uncertainty Quantification by Bayesian Inference [0.8388591755871735]
This study presents a Bayesian maximum textitaposteriori (MAP) framework for dynamical system identification from time-series data.
arXiv Detail & Related papers (2024-01-30T12:16:52Z) - Improving Sample Efficiency of Model-Free Algorithms for Zero-Sum Markov Games [66.2085181793014]
We show that a model-free stage-based Q-learning algorithm can enjoy the same optimality in the $H$ dependence as model-based algorithms.
Our algorithm features a key novel design of updating the reference value functions as the pair of optimistic and pessimistic value functions.
arXiv Detail & Related papers (2023-08-17T08:34:58Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Analysis of One-Hidden-Layer Neural Networks via the Resolvent Method [0.0]
Motivated by random neural networks, we consider the random matrix $M = Y Yast$ with $Y = f(WX)$.
We prove that the Stieltjes transform of the limiting spectral distribution satisfies a quartic self-consistent equation up to some error terms.
In addition, we extend the previous results to the case of additive bias $Y=f(WX+B)$ with $B$ being an independent rank-one Gaussian random matrix.
arXiv Detail & Related papers (2021-05-11T15:17:39Z) - Self-supervised Symmetric Nonnegative Matrix Factorization [82.59905231819685]
Symmetric nonnegative factor matrix (SNMF) has demonstrated to be a powerful method for data clustering.
Inspired by ensemble clustering that aims to seek better clustering results, we propose self-supervised SNMF (S$3$NMF)
We take advantage of the sensitivity to code characteristic of SNMF, without relying on any additional information.
arXiv Detail & Related papers (2021-03-02T12:47:40Z) - Convergence of Langevin Monte Carlo in Chi-Squared and Renyi Divergence [8.873449722727026]
We show that the rate estimate $widetildemathcalO(depsilon-1)$ improves the previously known rates in both of these metrics.
In particular, for convex and firstorder smooth potentials, we show that LMC algorithm achieves the rate estimate $widetildemathcalO(depsilon-1)$ which improves the previously known rates in both of these metrics.
arXiv Detail & Related papers (2020-07-22T18:18:28Z) - Debiasing Distributed Second Order Optimization with Surrogate Sketching
and Scaled Regularization [101.5159744660701]
In distributed second order optimization, a standard strategy is to average many local estimates, each of which is based on a small sketch or batch of the data.
Here, we introduce a new technique for debiasing the local estimates, which leads to both theoretical and empirical improvements in the convergence rate of distributed second order methods.
arXiv Detail & Related papers (2020-07-02T18:08:14Z) - A Precise High-Dimensional Asymptotic Theory for Boosting and
Minimum-$\ell_1$-Norm Interpolated Classifiers [3.167685495996986]
This paper establishes a precise high-dimensional theory for boosting on separable data.
Under a class of statistical models, we provide an exact analysis of the universality error of boosting.
We also explicitly pin down the relation between the boosting test error and the optimal Bayes error.
arXiv Detail & Related papers (2020-02-05T00:24:53Z) - A unified framework for hard and soft clustering with regularized optimal transport [5.715859759904031]
In this paper, we formulate the problem of inferring a Finitelamblambdageq 0 from discrete data as an optimal transport problem with entropic regularization.
arXiv Detail & Related papers (2017-11-12T21:52:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.