Entropy Minimizing Matrix Factorization
- URL: http://arxiv.org/abs/2103.13487v1
- Date: Wed, 24 Mar 2021 21:08:43 GMT
- Title: Entropy Minimizing Matrix Factorization
- Authors: Mulin Chen and Xuelong Li
- Abstract summary: Nonnegative Matrix Factorization (NMF) is a widely-used data analysis technique, and has yielded impressive results in many real-world tasks.
In this study, an Entropy Minimizing Matrix Factorization framework (EMMF) is developed to tackle the above problem.
Considering that the outliers are usually much less than the normal samples, a new entropy loss function is established for matrix factorization.
- Score: 102.26446204624885
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Nonnegative Matrix Factorization (NMF) is a widely-used data analysis
technique, and has yielded impressive results in many real-world tasks.
Generally, existing NMF methods represent each sample with several centroids,
and find the optimal centroids by minimizing the sum of the approximation
errors. However, the outliers deviating from the normal data distribution may
have large residues, and then dominate the objective value seriously. In this
study, an Entropy Minimizing Matrix Factorization framework (EMMF) is developed
to tackle the above problem. Considering that the outliers are usually much
less than the normal samples, a new entropy loss function is established for
matrix factorization, which minimizes the entropy of the residue distribution
and allows a few samples to have large approximation errors. In this way, the
outliers do not affect the approximation of the normal samples. The
multiplicative updating rules for EMMF are also designed, and the convergence
is proved both theoretically and experimentally. In addition, a Graph
regularized version of EMMF (G-EMMF) is also presented to deal with the complex
data structure. Clustering results on various synthetic and real-world datasets
demonstrate the reasonableness of the proposed models, and the effectiveness is
also verified through the comparison with the state-of-the-arts.
Related papers
- Towards a Fairer Non-negative Matrix Factorization [6.069820038869034]
We investigate how Non-negative Matrix Factorization (NMF) can introduce bias in the representation of data groups.
We present an approach, called Fairer-NMF, that seeks to minimize the maximum reconstruction loss for different groups.
arXiv Detail & Related papers (2024-11-14T23:34:38Z) - Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Contaminated Images Recovery by Implementing Non-negative Matrix
Factorisation [0.0]
We theoretically examine the robustness of the traditional NMF, HCNMF, and L2,1-NMF algorithms and execute sets of experiments to demonstrate the robustness on ORL and Extended YaleB datasets.
Due to the computational cost of these approaches, our final models, such as the HCNMF and L2,1-NMF model, fail to converge within the parameters of this work.
arXiv Detail & Related papers (2022-11-08T13:50:27Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Unitary Approximate Message Passing for Matrix Factorization [90.84906091118084]
We consider matrix factorization (MF) with certain constraints, which finds wide applications in various areas.
We develop a Bayesian approach to MF with an efficient message passing implementation, called UAMPMF.
We show that UAMPMF significantly outperforms state-of-the-art algorithms in terms of recovery accuracy, robustness and computational complexity.
arXiv Detail & Related papers (2022-07-31T12:09:32Z) - Log-based Sparse Nonnegative Matrix Factorization for Data
Representation [55.72494900138061]
Nonnegative matrix factorization (NMF) has been widely studied in recent years due to its effectiveness in representing nonnegative data with parts-based representations.
We propose a new NMF method with log-norm imposed on the factor matrices to enhance the sparseness.
A novel column-wisely sparse norm, named $ell_2,log$-(pseudo) norm, is proposed to enhance the robustness of the proposed method.
arXiv Detail & Related papers (2022-04-22T11:38:10Z) - Sampling Approximately Low-Rank Ising Models: MCMC meets Variational
Methods [35.24886589614034]
We consider quadratic definite Ising models on the hypercube with a general interaction $J$.
Our general result implies the first time sampling algorithms for low-rank Ising models.
arXiv Detail & Related papers (2022-02-17T21:43:50Z) - Data embedding and prediction by sparse tropical matrix factorization [0.0]
We propose a method called Sparse Tropical Matrix Factorization (STMF) for the estimation of missing (unknown) values.
Tests on unique synthetic data showed that STMF approximation achieves a higher correlation than non-negative matrix factorization.
STMF is the first work that uses tropical semiring on sparse data.
arXiv Detail & Related papers (2020-12-09T18:09:17Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Efficient MCMC Sampling for Bayesian Matrix Factorization by Breaking
Posterior Symmetries [1.3858051019755282]
We propose a simple modification to the prior choice that provably breaks these symmetries and maintains/improves accuracy.
We show that using non-zero linearly independent prior means significantly lowers the autocorrelation of MCMC samples, and can also lead to lower reconstruction errors.
arXiv Detail & Related papers (2020-06-08T00:25:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.