Reducing Redundancy in the Bottleneck Representation of the Autoencoders
- URL: http://arxiv.org/abs/2202.04629v1
- Date: Wed, 9 Feb 2022 18:48:02 GMT
- Title: Reducing Redundancy in the Bottleneck Representation of the Autoencoders
- Authors: Firas Laakom, Jenni Raitoharju, Alexandros Iosifidis and Moncef
Gabbouj
- Abstract summary: Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
- Score: 98.78384185493624
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autoencoders are a type of unsupervised neural networks, which can be used to
solve various tasks, e.g., dimensionality reduction, image compression, and
image denoising. An AE has two goals: (i) compress the original input to a
low-dimensional space at the bottleneck of the network topology using an
encoder, (ii) reconstruct the input from the representation at the bottleneck
using a decoder. Both encoder and decoder are optimized jointly by minimizing a
distortion-based loss which implicitly forces the model to keep only those
variations of input data that are required to reconstruct the and to reduce
redundancies. In this paper, we propose a scheme to explicitly penalize feature
redundancies in the bottleneck representation. To this end, we propose an
additional loss term, based on the pair-wise correlation of the neurons, which
complements the standard reconstruction loss forcing the encoder to learn a
more diverse and richer representation of the input. We tested our approach
across different tasks: dimensionality reduction using three different dataset,
image compression using the MNIST dataset, and image denoising using fashion
MNIST. The experimental results show that the proposed loss leads consistently
to superior performance compared to the standard AE loss.
Related papers
- $ε$-VAE: Denoising as Visual Decoding [61.29255979767292]
In generative modeling, tokenization simplifies complex data into compact, structured representations, creating a more efficient, learnable space.
Current visual tokenization methods rely on a traditional autoencoder framework, where the encoder compresses data into latent representations, and the decoder reconstructs the original input.
We propose denoising as decoding, shifting from single-step reconstruction to iterative refinement. Specifically, we replace the decoder with a diffusion process that iteratively refines noise to recover the original image, guided by the latents provided by the encoder.
We evaluate our approach by assessing both reconstruction (rFID) and generation quality (
arXiv Detail & Related papers (2024-10-05T08:27:53Z) - Compression of Structured Data with Autoencoders: Provable Benefit of
Nonlinearities and Depth [83.15263499262824]
We prove that gradient descent converges to a solution that completely disregards the sparse structure of the input.
We show how to improve upon Gaussian performance for the compression of sparse data by adding a denoising function to a shallow architecture.
We validate our findings on image datasets, such as CIFAR-10 and MNIST.
arXiv Detail & Related papers (2024-02-07T16:32:29Z) - Compression with Bayesian Implicit Neural Representations [16.593537431810237]
We propose overfitting variational neural networks to the data and compressing an approximate posterior weight sample using relative entropy coding instead of quantizing and entropy coding it.
Experiments show that our method achieves strong performance on image and audio compression while retaining simplicity.
arXiv Detail & Related papers (2023-05-30T16:29:52Z) - Modeling Image Quantization Tradeoffs for Optimal Compression [0.0]
Lossy compression algorithms target tradeoffs by quantizating high frequency data to increase compression rates.
We propose a new method of optimizing quantization tables using Deep Learning and a minimax loss function.
arXiv Detail & Related papers (2021-12-14T07:35:22Z) - Small Lesion Segmentation in Brain MRIs with Subpixel Embedding [105.1223735549524]
We present a method to segment MRI scans of the human brain into ischemic stroke lesion and normal tissues.
We propose a neural network architecture in the form of a standard encoder-decoder where predictions are guided by a spatial expansion embedding network.
arXiv Detail & Related papers (2021-09-18T00:21:17Z) - Dynamic Neural Representational Decoders for High-Resolution Semantic
Segmentation [98.05643473345474]
We propose a novel decoder, termed dynamic neural representational decoder (NRD)
As each location on the encoder's output corresponds to a local patch of the semantic labels, in this work, we represent these local patches of labels with compact neural networks.
This neural representation enables our decoder to leverage the smoothness prior in the semantic label space, and thus makes our decoder more efficient.
arXiv Detail & Related papers (2021-07-30T04:50:56Z) - Orthogonal Features-based EEG Signal Denoising using Fractionally
Compressed AutoEncoder [16.889633963766858]
A fractional-based compressed auto-encoder architecture has been introduced to solve the problem of denoising electroencephalogram (EEG) signals.
The proposed architecture provides improved denoising results on the standard datasets when compared with the existing methods.
arXiv Detail & Related papers (2021-02-16T11:15:00Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.