An Empirical Analysis of Recurrent Learning Algorithms In Neural Lossy
Image Compression Systems
- URL: http://arxiv.org/abs/2201.11782v1
- Date: Thu, 27 Jan 2022 19:47:51 GMT
- Title: An Empirical Analysis of Recurrent Learning Algorithms In Neural Lossy
Image Compression Systems
- Authors: Ankur Mali and Alexander Ororbia and Daniel Kifer and Lee Giles
- Abstract summary: Recent advances in deep learning have resulted in image compression algorithms that outperform JPEG and JPEG 2000 on the standard Kodak benchmark.
In this paper, we perform the first large-scale comparison of recent state-of-the-art hybrid neural compression algorithms.
- Score: 73.48927855855219
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent advances in deep learning have resulted in image compression
algorithms that outperform JPEG and JPEG 2000 on the standard Kodak benchmark.
However, they are slow to train (due to backprop-through-time) and, to the best
of our knowledge, have not been systematically evaluated on a large variety of
datasets. In this paper, we perform the first large-scale comparison of recent
state-of-the-art hybrid neural compression algorithms, while exploring the
effects of alternative training strategies (when applicable). The hybrid
recurrent neural decoder is a former state-of-the-art model (recently overtaken
by a Google model) that can be trained using backprop-through-time (BPTT) or
with alternative algorithms like sparse attentive backtracking (SAB), unbiased
online recurrent optimization (UORO), and real-time recurrent learning (RTRL).
We compare these training alternatives along with the Google models (GOOG and
E2E) on 6 benchmark datasets. Surprisingly, we found that the model trained
with SAB performs better (outperforming even BPTT), resulting in faster
convergence and a better peak signal-to-noise ratio.
Related papers
- LeRF: Learning Resampling Function for Adaptive and Efficient Image Interpolation [64.34935748707673]
Recent deep neural networks (DNNs) have made impressive progress in performance by introducing learned data priors.
We propose a novel method of Learning Resampling (termed LeRF) which takes advantage of both the structural priors learned by DNNs and the locally continuous assumption.
LeRF assigns spatially varying resampling functions to input image pixels and learns to predict the shapes of these resampling functions with a neural network.
arXiv Detail & Related papers (2024-07-13T16:09:45Z) - Bit Rate Matching Algorithm Optimization in JPEG-AI Verification Model [45.64945099984195]
The research on neural network (NN) based image compression has shown superior performance compared to classical compression frameworks.
The current state of the JPEG-AI verification model experiences significant slowdowns during bit rate matching.
The proposed methodology offers a fourfold acceleration and over 1% improvement in BD-rate at the base operation point.
arXiv Detail & Related papers (2024-02-27T13:12:18Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Loop Unrolled Shallow Equilibrium Regularizer (LUSER) -- A
Memory-Efficient Inverse Problem Solver [26.87738024952936]
In inverse problems we aim to reconstruct some underlying signal of interest from potentially corrupted and often ill-posed measurements.
We propose an LU algorithm with shallow equilibrium regularizers (L)
These implicit models are as expressive as deeper convolutional networks, but far more memory efficient during training.
arXiv Detail & Related papers (2022-10-10T19:50:37Z) - Image reconstruction algorithms in radio interferometry: from
handcrafted to learned denoisers [7.1439425093981574]
We introduce a new class of iterative image reconstruction algorithms for radio interferometry, inspired by plug-and-play methods.
The approach consists in learning a prior image model by training a deep neural network (DNN) as a denoiser.
We plug the learned denoiser into the forward-backward optimization algorithm, resulting in a simple iterative structure alternating a denoising step with a gradient-descent data-fidelity step.
arXiv Detail & Related papers (2022-02-25T20:26:33Z) - A new Sparse Auto-encoder based Framework using Grey Wolf Optimizer for
Data Classification Problem [0.0]
Gray wolf optimization (GWO) is applied to train sparse auto-encoders.
Model is validated by employing several popular Gene expression databases.
Results reveal that the performance of the trained model using GWO outperforms on both conventional models and models trained with most popular metaheuristic algorithms.
arXiv Detail & Related papers (2022-01-29T04:28:30Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork [73.94896986868146]
Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
arXiv Detail & Related papers (2021-01-12T08:36:23Z) - Neural Network-based Reconstruction in Compressed Sensing MRI Without
Fully-sampled Training Data [17.415937218905125]
CS-MRI has shown promise in reconstructing under-sampled MR images.
Deep learning models have been developed that model the iterative nature of classical techniques by unrolling iterations in a neural network.
In this paper, we explore a novel strategy to train an unrolled reconstruction network in an unsupervised fashion by adopting a loss function widely-used in classical optimization schemes.
arXiv Detail & Related papers (2020-07-29T17:46:55Z) - BP-DIP: A Backprojection based Deep Image Prior [49.375539602228415]
We propose two image restoration approaches: (i) Deep Image Prior (DIP), which trains a convolutional neural network (CNN) from scratch in test time using the degraded image; and (ii) a backprojection (BP) fidelity term, which is an alternative to the standard least squares loss that is usually used in previous DIP works.
We demonstrate the performance of the proposed method, termed BP-DIP, on the deblurring task and show its advantages over the plain DIP, with both higher PSNR values and better inference run-time.
arXiv Detail & Related papers (2020-03-11T17:09:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.