Cascade Decoders-Based Autoencoders for Image Reconstruction
- URL: http://arxiv.org/abs/2107.00002v1
- Date: Tue, 29 Jun 2021 23:40:54 GMT
- Title: Cascade Decoders-Based Autoencoders for Image Reconstruction
- Authors: Honggui Li, Dimitri Galayko, Maria Trocan, Mohamad Sawan
- Abstract summary: This paper aims for image reconstruction of autoencoders, employs cascade decoders-based autoencoders.
The proposed serial decoders-based autoencoders include the architectures of multi-level decoders and the related optimization algorithms.
It is evaluated by the experimental results that the proposed autoencoders outperform the classical autoencoders in the performance of image reconstruction.
- Score: 2.924868086534434
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autoencoders are composed of coding and decoding units, hence they hold the
inherent potential of high-performance data compression and signal compressed
sensing. The main disadvantages of current autoencoders comprise the following
several aspects: the research objective is not data reconstruction but feature
representation; the performance evaluation of data recovery is neglected; it is
hard to achieve lossless data reconstruction by pure autoencoders, even by pure
deep learning. This paper aims for image reconstruction of autoencoders,
employs cascade decoders-based autoencoders, perfects the performance of image
reconstruction, approaches gradually lossless image recovery, and provides
solid theory and application basis for autoencoders-based image compression and
compressed sensing. The proposed serial decoders-based autoencoders include the
architectures of multi-level decoders and the related optimization algorithms.
The cascade decoders consist of general decoders, residual decoders,
adversarial decoders and their combinations. It is evaluated by the
experimental results that the proposed autoencoders outperform the classical
autoencoders in the performance of image reconstruction.
Related papers
- $ε$-VAE: Denoising as Visual Decoding [61.29255979767292]
In generative modeling, tokenization simplifies complex data into compact, structured representations, creating a more efficient, learnable space.
Current visual tokenization methods rely on a traditional autoencoder framework, where the encoder compresses data into latent representations, and the decoder reconstructs the original input.
We propose denoising as decoding, shifting from single-step reconstruction to iterative refinement. Specifically, we replace the decoder with a diffusion process that iteratively refines noise to recover the original image, guided by the latents provided by the encoder.
We evaluate our approach by assessing both reconstruction (rFID) and generation quality (
arXiv Detail & Related papers (2024-10-05T08:27:53Z) - Neural Image Compression Using Masked Sparse Visual Representation [17.229601298529825]
We study neural image compression based on the Sparse Visual Representation (SVR), where images are embedded into a discrete latent space spanned by learned visual codebooks.
By sharing codebooks with the decoder, the encoder transfers codeword indices that are efficient and cross-platform robust.
We propose a Masked Adaptive Codebook learning (M-AdaCode) method that applies masks to the latent feature subspace to balance and reconstruction quality.
arXiv Detail & Related papers (2023-09-20T21:59:23Z) - VNVC: A Versatile Neural Video Coding Framework for Efficient
Human-Machine Vision [59.632286735304156]
It is more efficient to enhance/analyze the coded representations directly without decoding them into pixels.
We propose a versatile neural video coding (VNVC) framework, which targets learning compact representations to support both reconstruction and direct enhancement/analysis.
arXiv Detail & Related papers (2023-06-19T03:04:57Z) - Deep Lossy Plus Residual Coding for Lossless and Near-lossless Image
Compression [85.93207826513192]
We propose a unified and powerful deep lossy plus residual (DLPR) coding framework for both lossless and near-lossless image compression.
We solve the joint lossy and residual compression problem in the approach of VAEs.
In the near-lossless mode, we quantize the original residuals to satisfy a given $ell_infty$ error bound.
arXiv Detail & Related papers (2022-09-11T12:11:56Z) - On a Mechanism Framework of Autoencoders [0.0]
This paper proposes a theoretical framework on the mechanism of autoencoders.
Results of ReLU autoencoders are generalized to some non-ReLU cases.
Compared to PCA and decision trees, the advantages of (generalized) autoencoders on dimensionality reduction and classification are demonstrated.
arXiv Detail & Related papers (2022-08-15T03:51:40Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - HARP-Net: Hyper-Autoencoded Reconstruction Propagation for Scalable
Neural Audio Coding [25.51661602383911]
An autoencoder-based decoder employs quantization to turn its bottleneck layer activation into bitstrings.
To circumvent this issue, we employ additional skip connections between the corresponding pair of encoder-decoder layers.
We empirically verify that the proposed hyper-autoencoded architecture improves audio quality compared to an ordinary autoencoder baseline.
arXiv Detail & Related papers (2021-07-22T17:57:53Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.