Two to Five Truths in Non-Negative Matrix Factorization
- URL: http://arxiv.org/abs/2305.05389v2
- Date: Tue, 5 Sep 2023 16:14:56 GMT
- Title: Two to Five Truths in Non-Negative Matrix Factorization
- Authors: John M. Conroy, Neil P Molino, Brian Baughman, Rod Gomez, Ryan
Kaliszewski, and Nicholas A. Lines
- Abstract summary: We present a scaling inspired by the normalized Laplacian (NL) for graphs that can greatly improve the quality of a non-negative matrix factorization.
The matrix scaling gives significant improvement for text topic models in a variety of datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we explore the role of matrix scaling on a matrix of counts
when building a topic model using non-negative matrix factorization. We present
a scaling inspired by the normalized Laplacian (NL) for graphs that can greatly
improve the quality of a non-negative matrix factorization. The results
parallel those in the spectral graph clustering work of \cite{Priebe:2019},
where the authors proved adjacency spectral embedding (ASE) spectral clustering
was more likely to discover core-periphery partitions and Laplacian Spectral
Embedding (LSE) was more likely to discover affinity partitions. In text
analysis non-negative matrix factorization (NMF) is typically used on a matrix
of co-occurrence ``contexts'' and ``terms" counts. The matrix scaling inspired
by LSE gives significant improvement for text topic models in a variety of
datasets. We illustrate the dramatic difference a matrix scalings in NMF can
greatly improve the quality of a topic model on three datasets where human
annotation is available. Using the adjusted Rand index (ARI), a measure cluster
similarity we see an increase of 50\% for Twitter data and over 200\% for a
newsgroup dataset versus using counts, which is the analogue of ASE. For clean
data, such as those from the Document Understanding Conference, NL gives over
40\% improvement over ASE. We conclude with some analysis of this phenomenon
and some connections of this scaling with other matrix scaling methods.
Related papers
- Understanding Matrix Function Normalizations in Covariance Pooling through the Lens of Riemannian Geometry [63.694184882697435]
Global Covariance Pooling (GCP) has been demonstrated to improve the performance of Deep Neural Networks (DNNs) by exploiting second-order statistics of high-level representations.
arXiv Detail & Related papers (2024-07-15T07:11:44Z) - Interpretable Multi-View Clustering Based on Anchor Graph Tensor Factorization [64.00146569922028]
Multi-view clustering methods based on anchor graph factorization lack adequate cluster interpretability for the decomposed matrix.
We address this limitation by using non-negative tensor factorization to decompose an anchor graph tensor that combines anchor graphs from multiple views.
arXiv Detail & Related papers (2024-04-01T03:23:55Z) - Clustering Three-Way Data with Outliers [1.0435741631709405]
An approach for clustering matrix-variate normal data with outliers is discussed.
The approach, which uses the distribution of subset log-likelihoods, extends the OCLUST algorithm and uses an iterative approach to detect and trim outliers.
arXiv Detail & Related papers (2023-10-08T21:27:29Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Stochastic Parallelizable Eigengap Dilation for Large Graph Clustering [12.544602297450533]
We aim to identify clusters of nodes where most edges fall within clusters and only few edges fall between clusters.
A core step of spectral clustering is performing an eigendecomposition of the corresponding graph Laplacian matrix.
This paper introduces a parallelizable approach to dilating the spectrum in order to accelerate SVD solvers and in turn, spectral clustering.
arXiv Detail & Related papers (2022-07-29T10:13:07Z) - Optimal Variable Clustering for High-Dimensional Matrix Valued Data [3.1138411427556445]
We propose a new latent variable model for the features arranged in matrix form.
Under mild conditions, our algorithm attains clustering consistency in the high-dimensional setting.
We identify the optimal weight in the sense that using this weight guarantees our algorithm to be minimax rate-optimal.
arXiv Detail & Related papers (2021-12-24T02:13:04Z) - Dictionary-based Low-Rank Approximations and the Mixed Sparse Coding
problem [7.132368785057316]
I show how to adapt an efficient MSC solver based on the LASSO to compute Dictionary-based Matrix Factorization and Canonical Polyadic Decomposition.
I show how to adapt an efficient MSC solver based on the LASSO to compute Dictionary-based Matrix Factorization and Canonical Polyadic Decomposition in the context of hyperspectral image processing and chemometrics.
arXiv Detail & Related papers (2021-11-24T10:32:48Z) - Sawtooth Factorial Topic Embeddings Guided Gamma Belief Network [49.458250193768826]
We propose sawtooth factorial topic embedding guided GBN, a deep generative model of documents.
Both the words and topics are represented as embedding vectors of the same dimension.
Our models outperform other neural topic models on extracting deeper interpretable topics.
arXiv Detail & Related papers (2021-06-30T10:14:57Z) - Feature Weighted Non-negative Matrix Factorization [92.45013716097753]
We propose the Feature weighted Non-negative Matrix Factorization (FNMF) in this paper.
FNMF learns the weights of features adaptively according to their importances.
It can be solved efficiently with the suggested optimization algorithm.
arXiv Detail & Related papers (2021-03-24T21:17:17Z) - Supervised Quantile Normalization for Low-rank Matrix Approximation [50.445371939523305]
We learn the parameters of quantile normalization operators that can operate row-wise on the values of $X$ and/or of its factorization $UV$ to improve the quality of the low-rank representation of $X$ itself.
We demonstrate the applicability of these techniques on synthetic and genomics datasets.
arXiv Detail & Related papers (2020-02-08T21:06:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.