Sparse Coding with Multi-Layer Decoders using Variance Regularization
- URL: http://arxiv.org/abs/2112.09214v1
- Date: Thu, 16 Dec 2021 21:46:23 GMT
- Title: Sparse Coding with Multi-Layer Decoders using Variance Regularization
- Authors: Katrina Evtimova, Yann LeCun
- Abstract summary: We propose a novel sparse coding protocol which prevents a collapse in the codes without the need to regularize the decoder.
Our method regularizes the codes directly so that each latent code component has variance greater than a fixed threshold.
We show that sparse autoencoders with multi-layer decoders trained using our variance regularization method produce higher quality reconstructions with sparser representations.
- Score: 19.8572592390623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sparse coding with an $l_1$ penalty and a learned linear dictionary requires
regularization of the dictionary to prevent a collapse in the $l_1$ norms of
the codes. Typically, this regularization entails bounding the Euclidean norms
of the dictionary's elements. In this work, we propose a novel sparse coding
protocol which prevents a collapse in the codes without the need to regularize
the decoder. Our method regularizes the codes directly so that each latent code
component has variance greater than a fixed threshold over a set of sparse
representations for a given set of inputs. Furthermore, we explore ways to
effectively train sparse coding systems with multi-layer decoders since they
can model more complex relationships than linear dictionaries. In our
experiments with MNIST and natural image patches, we show that decoders learned
with our approach have interpretable features both in the linear and
multi-layer case. Moreover, we show that sparse autoencoders with multi-layer
decoders trained using our variance regularization method produce higher
quality reconstructions with sparser representations when compared to
autoencoders with linear dictionaries. Additionally, sparse representations
obtained with our variance regularization approach are useful in the downstream
tasks of denoising and classification in the low-data regime.
Related papers
- Triple-Encoders: Representations That Fire Together, Wire Together [51.15206713482718]
Contrastive Learning is a representation learning method that encodes relative distances between utterances into the embedding space via a bi-encoder.
This study introduces triple-encoders, which efficiently compute distributed utterance mixtures from these independently encoded utterances.
We find that triple-encoders lead to a substantial improvement over bi-encoders, and even to better zero-shot generalization than single-vector representation models.
arXiv Detail & Related papers (2024-02-19T18:06:02Z) - SparseCoder: Identifier-Aware Sparse Transformer for File-Level Code
Summarization [51.67317895094664]
This paper studies file-level code summarization, which can assist programmers in understanding and maintaining large source code projects.
We propose SparseCoder, an identifier-aware sparse transformer for effectively handling long code sequences.
arXiv Detail & Related papers (2024-01-26T09:23:27Z) - Deep Polar Codes [19.265010348250897]
We introduce a novel class of pre-transformed polar codes, termed as deep polar codes.
We first present a deep polar encoder that harnesses a series of multi-layered polar transformations with varying sizes.
Our encoding method offers flexibility in rate-profiling, embracing a wide range of code rates and blocklengths.
arXiv Detail & Related papers (2023-08-06T03:29:18Z) - SC-VAE: Sparse Coding-based Variational Autoencoder with Learned ISTA [0.6770292596301478]
We introduce a new VAE variant, termed sparse coding-based VAE with learned ISTA (SC-VAE), which integrates sparse coding within variational autoencoder framework.
Experiments on two image datasets demonstrate that our model achieves improved image reconstruction results compared to state-of-the-art methods.
arXiv Detail & Related papers (2023-03-29T13:18:33Z) - Variational Sparse Coding with Learned Thresholding [6.737133300781134]
We propose a new approach to variational sparse coding that allows us to learn sparse distributions by thresholding samples.
We first evaluate and analyze our method by training a linear generator, showing that it has superior performance, statistical efficiency, and gradient estimation.
arXiv Detail & Related papers (2022-05-07T14:49:50Z) - Sensitivity of sparse codes to image distortions [4.209801809583906]
We show that sparse codes can be very sensitive to image distortions.
The sensitivity is due to the existence of linear combinations of active dictionary elements with high cancellation.
arXiv Detail & Related papers (2022-04-15T13:58:00Z) - Hierarchical Sketch Induction for Paraphrase Generation [79.87892048285819]
We introduce Hierarchical Refinement Quantized Variational Autoencoders (HRQ-VAE), a method for learning decompositions of dense encodings.
We use HRQ-VAE to encode the syntactic form of an input sentence as a path through the hierarchy, allowing us to more easily predict syntactic sketches at test time.
arXiv Detail & Related papers (2022-03-07T15:28:36Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - The Interpretable Dictionary in Sparse Coding [4.205692673448206]
In our work, we illustrate that an ANN, trained using sparse coding under specific sparsity constraints, yields a more interpretable model than the standard deep learning model.
The dictionary learned by sparse coding can be more easily understood and the activations of these elements creates a selective feature output.
arXiv Detail & Related papers (2020-11-24T00:26:40Z) - Learning Autoencoders with Relational Regularization [89.53065887608088]
A new framework is proposed for learning autoencoders of data distributions.
We minimize the discrepancy between the model and target distributions, with a emphrelational regularization
We implement the framework with two scalable algorithms, making it applicable for both probabilistic and deterministic autoencoders.
arXiv Detail & Related papers (2020-02-07T17:27:30Z) - Pruning Neural Belief Propagation Decoders [77.237958592189]
We introduce a method to tailor an overcomplete parity-check matrix to (neural) BP decoding using machine learning.
We achieve performance within 0.27 dB and 1.5 dB of the ML performance while reducing the complexity of the decoder.
arXiv Detail & Related papers (2020-01-21T12:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.