Semi-supervised Nonnegative Matrix Factorization for Document
Classification
- URL: http://arxiv.org/abs/2203.03551v1
- Date: Mon, 28 Feb 2022 19:00:49 GMT
- Title: Semi-supervised Nonnegative Matrix Factorization for Document
Classification
- Authors: Jamie Haddock, Lara Kassab, Sixian Li, Alona Kryshchenko, Rachel
Grotheer, Elena Sizikova, Chuntian Wang, Thomas Merkh, RWMA Madushani, Miju
Ahn, Deanna Needell, Kathryn Leonard
- Abstract summary: We propose new semi-supervised nonnegative matrix factorization (SSNMF) models for document classification.
We derive training methods using multiplicative updates for each new model, and demonstrate the application of these models to single-label and multi-label document classification.
- Score: 6.577559557980527
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose new semi-supervised nonnegative matrix factorization (SSNMF)
models for document classification and provide motivation for these models as
maximum likelihood estimators. The proposed SSNMF models simultaneously provide
both a topic model and a model for classification, thereby offering highly
interpretable classification results. We derive training methods using
multiplicative updates for each new model, and demonstrate the application of
these models to single-label and multi-label document classification, although
the models are flexible to other supervised learning tasks such as regression.
We illustrate the promise of these models and training methods on document
classification datasets (e.g., 20 Newsgroups, Reuters).
Related papers
- Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Exploring Beyond Logits: Hierarchical Dynamic Labeling Based on Embeddings for Semi-Supervised Classification [49.09505771145326]
We propose a Hierarchical Dynamic Labeling (HDL) algorithm that does not depend on model predictions and utilizes image embeddings to generate sample labels.
Our approach has the potential to change the paradigm of pseudo-label generation in semi-supervised learning.
arXiv Detail & Related papers (2024-04-26T06:00:27Z) - Representer Point Selection for Explaining Regularized High-dimensional
Models [105.75758452952357]
We introduce a class of sample-based explanations we term high-dimensional representers.
Our workhorse is a novel representer theorem for general regularized high-dimensional models.
We study the empirical performance of our proposed methods on three real-world binary classification datasets and two recommender system datasets.
arXiv Detail & Related papers (2023-05-31T16:23:58Z) - Training Experimentally Robust and Interpretable Binarized Regression
Models Using Mixed-Integer Programming [3.179831861897336]
We present a model-based approach to training robust and interpretable binarized regression models for multiclass classification tasks.
Our MIP model balances the optimization of prediction margin and model size by using a weighted objective.
We show the effectiveness of training robust and interpretable binarized regression models using MIP.
arXiv Detail & Related papers (2021-12-01T11:53:08Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - On a Guided Nonnegative Matrix Factorization [9.813862201223973]
We propose an approach based upon the nonnegative matrix factorization (NMF) model, deemed textit NMF, that incorporates user-designed seed word supervision.
Our experimental results demonstrate the promise of this model and illustrate that it is competitive with other methods of this ilk with only very little supervision information.
arXiv Detail & Related papers (2020-10-22T01:06:17Z) - Semi-supervised NMF Models for Topic Modeling in Learning Tasks [6.577559557980527]
We propose several new models for semi-supervised nonnegative matrix factorization (SSNMF)
We present multiplicative updates training methods for each new model, and demonstrate the application of these models to classification.
arXiv Detail & Related papers (2020-10-15T18:03:46Z) - Robust Finite Mixture Regression for Heterogeneous Targets [70.19798470463378]
We propose an FMR model that finds sample clusters and jointly models multiple incomplete mixed-type targets simultaneously.
We provide non-asymptotic oracle performance bounds for our model under a high-dimensional learning framework.
The results show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-10-12T03:27:07Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.