A variational autoencoder-based nonnegative matrix factorisation model
for deep dictionary learning
- URL: http://arxiv.org/abs/2301.07272v1
- Date: Wed, 18 Jan 2023 02:36:03 GMT
- Title: A variational autoencoder-based nonnegative matrix factorisation model
for deep dictionary learning
- Authors: Hong-Bo Xie, Caoyuan Li, Shuliang Wang, Richard Yi Da Xu and Kerrie
Mengersen
- Abstract summary: Construction of dictionaries using nonnegative matrix factorisation (NMF) has extensive applications in signal processing and machine learning.
We propose a probabilistic generative model which employs a variational autoencoder (VAE) to perform nonnegative dictionary learning.
- Score: 13.796655751448288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Construction of dictionaries using nonnegative matrix factorisation (NMF) has
extensive applications in signal processing and machine learning. With the
advances in deep learning, training compact and robust dictionaries using deep
neural networks, i.e., dictionaries of deep features, has been proposed. In
this study, we propose a probabilistic generative model which employs a
variational autoencoder (VAE) to perform nonnegative dictionary learning. In
contrast to the existing VAE models, we cast the model under a statistical
framework with latent variables obeying a Gamma distribution and design a new
loss function to guarantee the nonnegative dictionaries. We adopt an
acceptance-rejection sampling reparameterization trick to update the latent
variables iteratively. We apply the dictionaries learned from VAE-NMF to two
signal processing tasks, i.e., enhancement of speech and extraction of muscle
synergies. Experimental results demonstrate that VAE-NMF performs better in
learning the latent nonnegative dictionaries in comparison with
state-of-the-art methods.
Related papers
- A Sparsity-promoting Dictionary Model for Variational Autoencoders [16.61511959679188]
Structuring the latent space in deep generative models is important to yield more expressive models and interpretable representations.
We propose a simple yet effective methodology to structure the latent space via a sparsity-promoting dictionary model.
arXiv Detail & Related papers (2022-03-29T17:13:11Z) - Better Language Model with Hypernym Class Prediction [101.8517004687825]
Class-based language models (LMs) have been long devised to address context sparsity in $n$-gram LMs.
In this study, we revisit this approach in the context of neural LMs.
arXiv Detail & Related papers (2022-03-21T01:16:44Z) - Distributionally Robust Recurrent Decoders with Random Network
Distillation [93.10261573696788]
We propose a method based on OOD detection with Random Network Distillation to allow an autoregressive language model to disregard OOD context during inference.
We apply our method to a GRU architecture, demonstrating improvements on multiple language modeling (LM) datasets.
arXiv Detail & Related papers (2021-10-25T19:26:29Z) - PUDLE: Implicit Acceleration of Dictionary Learning by Backpropagation [4.081440927534577]
This paper offers the first theoretical proof for empirical results through PUDLE, a Provable Unfolded Dictionary LEarning method.
We highlight the minimization impact of loss, unfolding, and backpropagation on convergence.
We complement our findings through synthetic and image denoising experiments.
arXiv Detail & Related papers (2021-05-31T18:49:58Z) - Defense against Adversarial Attacks in NLP via Dirichlet Neighborhood
Ensemble [163.3333439344695]
Dirichlet Neighborhood Ensemble (DNE) is a randomized smoothing method for training a robust model to defense substitution-based attacks.
DNE forms virtual sentences by sampling embedding vectors for each word in an input sentence from a convex hull spanned by the word and its synonyms, and it augments them with the training data.
We demonstrate through extensive experimentation that our method consistently outperforms recently proposed defense methods by a significant margin across different network architectures and multiple data sets.
arXiv Detail & Related papers (2020-06-20T18:01:16Z) - A Convolutional Deep Markov Model for Unsupervised Speech Representation
Learning [32.59760685342343]
Probabilistic Latent Variable Models provide an alternative to self-supervised learning approaches for linguistic representation learning from speech.
In this work, we propose ConvDMM, a Gaussian state-space model with non-linear emission and transition functions modelled by deep neural networks.
When trained on a large scale speech dataset (LibriSpeech), ConvDMM produces features that significantly outperform multiple self-supervised feature extracting methods.
arXiv Detail & Related papers (2020-06-03T21:50:20Z) - A Comparative Study of Lexical Substitution Approaches based on Neural
Language Models [117.96628873753123]
We present a large-scale comparative study of popular neural language and masked language models.
We show that already competitive results achieved by SOTA LMs/MLMs can be further improved if information about the target word is injected properly.
arXiv Detail & Related papers (2020-05-29T18:43:22Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z) - Contrast-weighted Dictionary Learning Based Saliency Detection for
Remote Sensing Images [3.338193485961624]
We propose a novel saliency detection model based on Contrast-weighted Dictionary Learning (CDL) for remote sensing images.
Specifically, the proposed CDL learns salient and non-salient atoms from positive and negative samples to construct a discriminant dictionary.
By using the proposed joint saliency measure, a variety of saliency maps are generated based on the discriminant dictionary.
arXiv Detail & Related papers (2020-04-06T06:49:05Z) - Learning Deep Analysis Dictionaries for Image Super-Resolution [38.7315182732103]
Deep Analysis dictionary Model (DeepAM) is optimized to address a specific regression task known as single image super-resolution.
Our architecture contains L layers of analysis dictionary and soft-thresholding operators.
DeepAM uses both supervised and unsupervised setup.
arXiv Detail & Related papers (2020-01-31T18:59:35Z) - Parameter Space Factorization for Zero-Shot Learning across Tasks and
Languages [112.65994041398481]
We propose a Bayesian generative model for the space of neural parameters.
We infer the posteriors over such latent variables based on data from seen task-language combinations.
Our model yields comparable or better results than state-of-the-art, zero-shot cross-lingual transfer methods.
arXiv Detail & Related papers (2020-01-30T16:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.