SUR-FeatNet: Predicting the Satisfied User Ratio Curvefor Image
Compression with Deep Feature Learning
- URL: http://arxiv.org/abs/2001.02002v2
- Date: Sun, 5 Apr 2020 08:55:22 GMT
- Title: SUR-FeatNet: Predicting the Satisfied User Ratio Curvefor Image
Compression with Deep Feature Learning
- Authors: Hanhe Lin, Vlad Hosu, Chunling Fan, Yun Zhang, Yuchen Mu, Raouf
Hamzaoui, Dietmar Saupe
- Abstract summary: We propose the first deep learning approach to predict SUR curves.
We show how to apply maximum likelihood estimation and the Anderson-Darling test to select a suitable parametric model.
Our deep learning approach relies on a siamese convolutional neural network, transfer learning, and deep feature learning.
- Score: 15.2348952809434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The satisfied user ratio (SUR) curve for a lossy image compression scheme,
e.g., JPEG, characterizes the complementary cumulative distribution function of
the just noticeable difference (JND), the smallest distortion level that can be
perceived by a subject when a reference image is compared to a distorted one. A
sequence of JNDs can be defined with a suitable successive choice of reference
images. We propose the first deep learning approach to predict SUR curves. We
show how to apply maximum likelihood estimation and the Anderson-Darling test
to select a suitable parametric model for the distribution function. We then
use deep feature learning to predict samples of the SUR curve and apply the
method of least squares to fit the parametric model to the predicted samples.
Our deep learning approach relies on a siamese convolutional neural network,
transfer learning, and deep feature learning, using pairs consisting of a
reference image and a compressed image for training. Experiments on the MCL-JCI
dataset showed state-of-the-art performance. For example, the mean
Bhattacharyya distances between the predicted and ground truth first, second,
and third JND distributions were 0.0810, 0.0702, and 0.0522, respectively, and
the corresponding average absolute differences of the peak signal-to-noise
ratio at a median of the first JND distribution were 0.58, 0.69, and 0.58 dB.
Further experiments on the JND-Pano dataset showed that the method transfers
well to high resolution panoramic images viewed on head-mounted displays.
Related papers
- Provably Robust Score-Based Diffusion Posterior Sampling for Plug-and-Play Image Reconstruction [31.503662384666274]
In science and engineering, the goal is to infer an unknown image from a small number of measurements collected from a known forward model describing certain imaging modality.
Motivated Score-based diffusion models, due to its empirical success, have emerged as an impressive candidate of an exemplary prior in image reconstruction.
arXiv Detail & Related papers (2024-03-25T15:58:26Z) - Ambient Diffusion Posterior Sampling: Solving Inverse Problems with
Diffusion Models trained on Corrupted Data [56.81246107125692]
Ambient Diffusion Posterior Sampling (A-DPS) is a generative model pre-trained on one type of corruption.
We show that A-DPS can sometimes outperform models trained on clean data for several image restoration tasks in both speed and performance.
We extend the Ambient Diffusion framework to train MRI models with access only to Fourier subsampled multi-coil MRI measurements.
arXiv Detail & Related papers (2024-03-13T17:28:20Z) - MB-RACS: Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network [65.1004435124796]
We propose a Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network (MB-RACS) framework.
Our experiments demonstrate that the proposed MB-RACS method surpasses current leading methods.
arXiv Detail & Related papers (2024-01-19T04:40:20Z) - B-Spine: Learning B-Spline Curve Representation for Robust and
Interpretable Spinal Curvature Estimation [50.208310028625284]
We propose B-Spine, a novel deep learning pipeline to learn B-spline curve representation of the spine.
We estimate the Cobb angles for spinal curvature estimation from low-quality X-ray images.
arXiv Detail & Related papers (2023-10-14T15:34:57Z) - Single Image Depth Prediction Made Better: A Multivariate Gaussian Take [163.14849753700682]
We introduce an approach that performs continuous modeling of per-pixel depth.
Our method's accuracy (named MG) is among the top on the KITTI depth-prediction benchmark leaderboard.
arXiv Detail & Related papers (2023-03-31T16:01:03Z) - Estimating the Resize Parameter in End-to-end Learned Image Compression [50.20567320015102]
We describe a search-free resizing framework that can further improve the rate-distortion tradeoff of recent learned image compression models.
Our results show that our new resizing parameter estimation framework can provide Bjontegaard-Delta rate (BD-rate) improvement of about 10% against leading perceptual quality engines.
arXiv Detail & Related papers (2022-04-26T01:35:02Z) - Score-based diffusion models for accelerated MRI [35.3148116010546]
We introduce a way to sample data from a conditional distribution given the measurements, such that the model can be readily used for solving inverse problems in imaging.
Our model requires magnitude images only for training, and yet is able to reconstruct complex-valued data, and even extends to parallel imaging.
arXiv Detail & Related papers (2021-10-08T08:42:03Z) - Accurate Cup-to-Disc Ratio Measurement with Tight Bounding Box
Supervision in Fundus Photography [5.517632401040172]
The cup-to-disc ratio (CDR) is one of the most significant indicator for glaucoma diagnosis.
This study investigates the feasibility of accurate CDR measurement in fundus images using only tight bounding box supervision.
arXiv Detail & Related papers (2021-10-03T07:23:40Z) - Multiscale Score Matching for Out-of-Distribution Detection [19.61640396236456]
We present a new methodology for detecting out-of-distribution (OOD) images by utilizing norms of the score estimates at multiple noise scales.
Our methodology is completely unsupervised and follows a straight forward training scheme.
arXiv Detail & Related papers (2020-10-25T15:10:31Z) - Learning Rates as a Function of Batch Size: A Random Matrix Theory
Approach to Neural Network Training [2.9649783577150837]
We study the effect of mini-batching on the loss landscape of deep neural networks using spiked, field-dependent random matrix theory.
We derive analytical expressions for the maximal descent and adaptive training regimens for smooth, non-Newton deep neural networks.
We validate our claims on the VGG/ResNet and ImageNet datasets.
arXiv Detail & Related papers (2020-06-16T11:55:45Z) - Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement [156.18634427704583]
The paper presents a novel method, Zero-Reference Deep Curve Estimation (Zero-DCE), which formulates light enhancement as a task of image-specific curve estimation with a deep network.
Our method trains a lightweight deep network, DCE-Net, to estimate pixel-wise and high-order curves for dynamic range adjustment of a given image.
arXiv Detail & Related papers (2020-01-19T13:49:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.