Neuromorphologicaly-preserving Volumetric data encoding using VQ-VAE
- URL: http://arxiv.org/abs/2002.05692v1
- Date: Thu, 13 Feb 2020 18:18:51 GMT
- Title: Neuromorphologicaly-preserving Volumetric data encoding using VQ-VAE
- Authors: Petru-Daniel Tudosiu and Thomas Varsavsky and Richard Shaw and Mark
Graham and Parashkev Nachev and Sebastien Ourselin and Carole H. Sudre and M.
Jorge Cardoso
- Abstract summary: We show a VQ-VAE inspired network can efficiently encode a full-resolution 3D brain volume, compressing the data to $0.825%$ of the original size while maintaining image fidelity.
We then demonstrate that VQ-VAE decoded images preserve the morphological characteristics of the original data through voxel-based morphology and segmentation experiments.
- Score: 4.221619479687068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The increasing efficiency and compactness of deep learning architectures,
together with hardware improvements, have enabled the complex and
high-dimensional modelling of medical volumetric data at higher resolutions.
Recently, Vector-Quantised Variational Autoencoders (VQ-VAE) have been proposed
as an efficient generative unsupervised learning approach that can encode
images to a small percentage of their initial size, while preserving their
decoded fidelity. Here, we show a VQ-VAE inspired network can efficiently
encode a full-resolution 3D brain volume, compressing the data to $0.825\%$ of
the original size while maintaining image fidelity, and significantly
outperforming the previous state-of-the-art. We then demonstrate that VQ-VAE
decoded images preserve the morphological characteristics of the original data
through voxel-based morphology and segmentation experiments. Lastly, we show
that such models can be pre-trained and then fine-tuned on different datasets
without the introduction of bias.
Related papers
- Quantum Down Sampling Filter for Variational Auto-encoder [0.504868948270058]
Variational Autoencoders (VAEs) are essential tools in generative modeling and image reconstruction.
This study aims to improve the quality of reconstructed images by enhancing their resolution and preserving finer details.
We propose a hybrid model that combines quantum computing techniques in the VAE encoder with convolutional neural networks (CNNs) in the decoder.
arXiv Detail & Related papers (2025-01-09T11:08:55Z) - Once-for-All: Controllable Generative Image Compression with Dynamic Granularity Adaption [52.82508784748278]
This paper proposes a Control Generative Image Compression framework, termed Control-GIC.
Control-GIC is capable of fine-grained adaption across a broad spectrum while ensuring high-fidelity and generality compression.
We develop a conditional decoder capable of retrieving historic multi-granularity representations according to encoded codes, and then reconstruct hierarchical features in the formalization of conditional probability.
arXiv Detail & Related papers (2024-06-02T14:22:09Z) - Efficient Visual State Space Model for Image Deblurring [83.57239834238035]
Convolutional neural networks (CNNs) and Vision Transformers (ViTs) have achieved excellent performance in image restoration.
We propose a simple yet effective visual state space model (EVSSM) for image deblurring.
arXiv Detail & Related papers (2024-05-23T09:13:36Z) - ConvNeXt-ChARM: ConvNeXt-based Transform for Efficient Neural Image
Compression [18.05997169440533]
We propose ConvNeXt-ChARM, an efficient ConvNeXt-based transform coding framework, paired with a compute-efficient channel-wise auto-regressive auto-regressive.
We show that ConvNeXt-ChARM brings consistent and significant BD-rate (PSNR) reductions estimated on average to 5.24% and 1.22% over the versatile video coding (VVC) reference encoder (VTM-18.0) and the state-of-the-art learned image compression method SwinT-ChARM.
arXiv Detail & Related papers (2023-07-12T11:45:54Z) - Joint Hierarchical Priors and Adaptive Spatial Resolution for Efficient
Neural Image Compression [11.25130799452367]
We propose an absolute image compression transformer (ICT) for neural image compression (NIC)
ICT captures both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents.
Our framework significantly improves the trade-off between coding efficiency and decoder complexity over the versatile video coding (VVC) reference encoder (VTM-18.0) and the neural SwinT-ChARM.
arXiv Detail & Related papers (2023-07-05T13:17:14Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - Hierarchical Residual Learning Based Vector Quantized Variational
Autoencoder for Image Reconstruction and Generation [19.92324010429006]
We propose a multi-layer variational autoencoder method, we call HR-VQVAE, that learns hierarchical discrete representations of the data.
We evaluate our method on the tasks of image reconstruction and generation.
arXiv Detail & Related papers (2022-08-09T06:04:25Z) - Neural Data-Dependent Transform for Learned Image Compression [72.86505042102155]
We build a neural data-dependent transform and introduce a continuous online mode decision mechanism to jointly optimize the coding efficiency for each individual image.
The experimental results show the effectiveness of the proposed neural-syntax design and the continuous online mode decision mechanism.
arXiv Detail & Related papers (2022-03-09T14:56:48Z) - Self-Supervised Variational Auto-Encoders [10.482805367361818]
We present a novel class of generative models, called self-supervised Variational Auto-Encoder (selfVAE)
This class of models allows to perform both conditional and unconditional sampling, while simplifying the objective function.
We present performance of our approach on three benchmark image data (Cifar10, Imagenette64, and CelebA)
arXiv Detail & Related papers (2020-10-05T13:42:28Z) - Hierarchical Amortized Training for Memory-efficient High Resolution 3D
GAN [52.851990439671475]
We propose a novel end-to-end GAN architecture that can generate high-resolution 3D images.
We achieve this goal by using different configurations between training and inference.
Experiments on 3D thorax CT and brain MRI demonstrate that our approach outperforms state of the art in image generation.
arXiv Detail & Related papers (2020-08-05T02:33:04Z) - NVAE: A Deep Hierarchical Variational Autoencoder [102.29977384039805]
We propose a deep hierarchical VAE built for image generation using depth-wise separable convolutions and batch normalization.
We show that NVAE achieves state-of-the-art results among non-autoregressive likelihood-based models.
To the best of our knowledge, NVAE is the first successful VAE applied to natural images as large as 256$times $256 pixels.
arXiv Detail & Related papers (2020-07-08T04:56:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.