Deep Optimized Multiple Description Image Coding via Scalar Quantization
Learning
- URL: http://arxiv.org/abs/2001.03851v1
- Date: Sun, 12 Jan 2020 05:03:16 GMT
- Title: Deep Optimized Multiple Description Image Coding via Scalar Quantization
Learning
- Authors: Lijun Zhao, Huihui Bai, Anhong Wang, Yao Zhao
- Abstract summary: We introduce a deep multiple description coding (MDC) framework optimized by minimizing multiple description (MD) compressive loss.
An auto-encoder network composed of these two types of network is designed as a symmetrical parameter sharing structure.
Our framework performs better than several state-of-the-art MDC approaches regarding image coding efficiency when tested on several commonly available datasets.
- Score: 37.00592782976494
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: In this paper, we introduce a deep multiple description coding (MDC)
framework optimized by minimizing multiple description (MD) compressive loss.
First, MD multi-scale-dilated encoder network generates multiple description
tensors, which are discretized by scalar quantizers, while these quantized
tensors are decompressed by MD cascaded-ResBlock decoder networks. To greatly
reduce the total amount of artificial neural network parameters, an
auto-encoder network composed of these two types of network is designed as a
symmetrical parameter sharing structure. Second, this autoencoder network and a
pair of scalar quantizers are simultaneously learned in an end-to-end
self-supervised way. Third, considering the variation in the image spatial
distribution, each scalar quantizer is accompanied by an importance-indicator
map to generate MD tensors, rather than using direct quantization. Fourth, we
introduce the multiple description structural similarity distance loss, which
implicitly regularizes the diversified multiple description generations, to
explicitly supervise multiple description diversified decoding in addition to
MD reconstruction loss. Finally, we demonstrate that our MDC framework performs
better than several state-of-the-art MDC approaches regarding image coding
efficiency when tested on several commonly available datasets.
Related papers
- ESDMR-Net: A Lightweight Network With Expand-Squeeze and Dual Multiscale
Residual Connections for Medical Image Segmentation [7.921517156237902]
This paper presents an expand-squeeze dual multiscale residual network ( ESDMR-Net)
It is a fully convolutional network that is well-suited for resource-constrained computing hardware such as mobile devices.
We present experiments on seven datasets from five distinct examples of applications.
arXiv Detail & Related papers (2023-12-17T02:15:49Z) - M$^{2}$SNet: Multi-scale in Multi-scale Subtraction Network for Medical
Image Segmentation [73.10707675345253]
We propose a general multi-scale in multi-scale subtraction network (M$2$SNet) to finish diverse segmentation from medical image.
Our method performs favorably against most state-of-the-art methods under different evaluation metrics on eleven datasets of four different medical image segmentation tasks.
arXiv Detail & Related papers (2023-03-20T06:26:49Z) - Perceptual Video Coding for Machines via Satisfied Machine Ratio
Modeling [66.56355316611598]
Satisfied Machine Ratio (SMR) is a metric that evaluates the perceptual quality of compressed images and videos for machines.
SMR enables perceptual coding for machines and propels Video Coding for Machines from specificity to generality.
arXiv Detail & Related papers (2022-11-13T03:16:36Z) - String-based Molecule Generation via Multi-decoder VAE [56.465033997245776]
We investigate the problem of string-based molecular generation via variational autoencoders (VAEs)
We propose a simple, yet effective idea to improve the performance of VAE for the task.
In our experiments, the proposed VAE model particularly performs well for generating a sample from out-of-domain distribution.
arXiv Detail & Related papers (2022-08-23T03:56:30Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - UNETR: Transformers for 3D Medical Image Segmentation [8.59571749685388]
We introduce a novel architecture, dubbed as UNEt TRansformers (UNETR), that utilizes a pure transformer as the encoder to learn sequence representations of the input volume.
We have extensively validated the performance of our proposed model across different imaging modalities.
arXiv Detail & Related papers (2021-03-18T20:17:15Z) - Semantic Features Aided Multi-Scale Reconstruction of Inter-Modality
Magnetic Resonance Images [12.39341163725669]
We propose a novel deep network based solution to reconstruct T2W images from T1W images (T1WI) using an encoder-decoder architecture.
The proposed learning is aided with semantic features by using multi-channel input with intensity values and gradient of image in two directions.
arXiv Detail & Related papers (2020-06-22T19:53:50Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.