End-to-End Learned Block-Based Image Compression with Block-Level Masked
Convolutions and Asymptotic Closed Loop Training
- URL: http://arxiv.org/abs/2203.11686v1
- Date: Tue, 22 Mar 2022 13:01:59 GMT
- Title: End-to-End Learned Block-Based Image Compression with Block-Level Masked
Convolutions and Asymptotic Closed Loop Training
- Authors: Fatih Kamisli
- Abstract summary: This paper explores an alternative learned block-based image compression approach in which neither an explicit intra prediction neural network nor an explicit deblocking neural network is used.
The experimental results indicate competitive image compression performance.
- Score: 2.741266294612776
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learned image compression research has achieved state-of-the-art compression
performance with auto-encoder based neural network architectures, where the
image is mapped via convolutional neural networks (CNN) into a latent
representation that is quantized and processed again with CNN to obtain the
reconstructed image. CNN operate on entire input images. On the other hand,
traditional state-of-the-art image and video compression methods process images
with a block-by-block processing approach for various reasons. Very recently,
work on learned image compression with block based approaches have also
appeared, which use the auto-encoder architecture on large blocks of the input
image and introduce additional neural networks that perform intra/spatial
prediction and deblocking/post-processing functions. This paper explores an
alternative learned block-based image compression approach in which neither an
explicit intra prediction neural network nor an explicit deblocking neural
network is used. A single auto-encoder neural network with block-level masked
convolutions is used and the block size is much smaller (8x8). By using
block-level masked convolutions, each block is processed using reconstructed
neighboring left and upper blocks both at the encoder and decoder. Hence, the
mutual information between adjacent blocks is exploited during compression and
each block is reconstructed using neighboring blocks, resolving the need for
explicit intra prediction and deblocking neural networks. Since the explored
system is a closed loop system, a special optimization procedure, the
asymptotic closed loop design, is used with standard stochastic gradient
descent based training. The experimental results indicate competitive image
compression performance.
Related papers
- A Deep Learning-based Compression and Classification Technique for Whole
Slide Histopathology Images [0.31498833540989407]
We build an ensemble of neural networks that enables a compressive autoencoder in a supervised fashion to retain a denser and more meaningful representation of the input histology images.
We test the compressed images using transfer learning-based classifiers and show that they provide promising accuracy and classification performance.
arXiv Detail & Related papers (2023-05-11T22:20:05Z) - Convolutional Neural Network (CNN) to reduce construction loss in JPEG
compression caused by Discrete Fourier Transform (DFT) [0.0]
Convolutional Neural Networks (CNN) have received more attention than most other types of deep neural networks.
In this work, an effective image compression method is purposed using autoencoders.
arXiv Detail & Related papers (2022-08-26T12:46:16Z) - The Devil Is in the Details: Window-based Attention for Image
Compression [58.1577742463617]
Most existing learned image compression models are based on Convolutional Neural Networks (CNNs)
In this paper, we study the effects of multiple kinds of attention mechanisms for local features learning, then introduce a more straightforward yet effective window-based local attention block.
The proposed window-based attention is very flexible which could work as a plug-and-play component to enhance CNN and Transformer models.
arXiv Detail & Related papers (2022-03-16T07:55:49Z) - COIN++: Data Agnostic Neural Compression [55.27113889737545]
COIN++ is a neural compression framework that seamlessly handles a wide range of data modalities.
We demonstrate the effectiveness of our method by compressing various data modalities.
arXiv Detail & Related papers (2022-01-30T20:12:04Z) - Implicit Neural Video Compression [17.873088127087605]
We propose a method to compress full-resolution video sequences with implicit neural representations.
Each frame is represented as a neural network that maps coordinate positions to pixel values.
We use a separate implicit network to modulate the coordinate inputs, which enables efficient motion compensation between frames.
arXiv Detail & Related papers (2021-12-21T15:59:00Z) - Small Lesion Segmentation in Brain MRIs with Subpixel Embedding [105.1223735549524]
We present a method to segment MRI scans of the human brain into ischemic stroke lesion and normal tissues.
We propose a neural network architecture in the form of a standard encoder-decoder where predictions are guided by a spatial expansion embedding network.
arXiv Detail & Related papers (2021-09-18T00:21:17Z) - Image Compression with Recurrent Neural Network and Generalized Divisive
Normalization [3.0204520109309843]
Deep learning has gained huge attention from the research community and produced promising image reconstruction results.
Recent methods focused on developing deeper and more complex networks, which significantly increased network complexity.
In this paper, two effective novel blocks are developed: analysis and block synthesis that employs the convolution layer and Generalized Divisive Normalization (GDN) in the variable-rate encoder and decoder side.
arXiv Detail & Related papers (2021-09-05T05:31:55Z) - Learned Multi-Resolution Variable-Rate Image Compression with
Octave-based Residual Blocks [15.308823742699039]
We propose a new variable-rate image compression framework, which employs generalized octave convolutions (GoConv) and generalized octave transposed-convolutions (GoTConv)
To enable a single model to operate with different bit rates and to learn multi-rate image features, a new objective function is introduced.
Experimental results show that the proposed framework trained with variable-rate objective function outperforms the standard codecs such as H.265/HEVC-based BPG and state-of-the-art learning-based variable-rate methods.
arXiv Detail & Related papers (2020-12-31T06:26:56Z) - Unfolding Neural Networks for Compressive Multichannel Blind
Deconvolution [71.29848468762789]
We propose a learned-structured unfolding neural network for the problem of compressive sparse multichannel blind-deconvolution.
In this problem, each channel's measurements are given as convolution of a common source signal and sparse filter.
We demonstrate that our method is superior to classical structured compressive sparse multichannel blind-deconvolution methods in terms of accuracy and speed of sparse filter recovery.
arXiv Detail & Related papers (2020-10-22T02:34:33Z) - Neural Sparse Representation for Image Restoration [116.72107034624344]
Inspired by the robustness and efficiency of sparse coding based image restoration models, we investigate the sparsity of neurons in deep networks.
Our method structurally enforces sparsity constraints upon hidden neurons.
Experiments show that sparse representation is crucial in deep neural networks for multiple image restoration tasks.
arXiv Detail & Related papers (2020-06-08T05:15:17Z) - Pyramid Attention Networks for Image Restoration [124.34970277136061]
Self-similarity refers to the image prior widely used in image restoration algorithms.
Recent advanced deep convolutional neural network based methods for image restoration do not take full advantage of self-similarities.
We present a novel Pyramid Attention module for image restoration, which captures long-range feature correspondences from a multi-scale feature pyramid.
arXiv Detail & Related papers (2020-04-28T21:12:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.