Lost in Compression: the Impact of Lossy Image Compression on Variable
Size Object Detection within Infrared Imagery
- URL: http://arxiv.org/abs/2205.08002v1
- Date: Mon, 16 May 2022 21:54:32 GMT
- Title: Lost in Compression: the Impact of Lossy Image Compression on Variable
Size Object Detection within Infrared Imagery
- Authors: Neelanjan Bhowmik, Jack W. Barker, Yona Falinie A. Gaus, Toby P.
Breckon
- Abstract summary: Lossy image compression strategies allow for more efficient storage and transmission of data by encoding data to a reduced form.
This is essential enable training with larger datasets on less storage-equipped environments.
In this work, we apply the lossy JPEG compression method with six discrete levels of increasing compression 95, 75, 50, 15, 10, 5 to infrared band (thermal) imagery.
- Score: 11.135527192198092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lossy image compression strategies allow for more efficient storage and
transmission of data by encoding data to a reduced form. This is essential
enable training with larger datasets on less storage-equipped environments.
However, such compression can cause severe decline in performance of deep
Convolution Neural Network (CNN) architectures even when mild compression is
applied and the resulting compressed imagery is visually identical. In this
work, we apply the lossy JPEG compression method with six discrete levels of
increasing compression {95, 75, 50, 15, 10, 5} to infrared band (thermal)
imagery. Our study quantitatively evaluates the affect that increasing levels
of lossy compression has upon the performance of characteristically diverse
object detection architectures (Cascade-RCNN, FSAF and Deformable DETR) with
respect to varying sizes of objects present in the dataset. When training and
evaluating on uncompressed data as a baseline, we achieve maximal mean Average
Precision (mAP) of 0.823 with Cascade R-CNN across the FLIR dataset,
outperforming prior work. The impact of the lossy compression is more extreme
at higher compression levels (15, 10, 5) across all three CNN architectures.
However, re-training models on lossy compressed imagery notably ameliorated
performances for all three CNN models with an average increment of ~76% (at
higher compression level 5). Additionally, we demonstrate the relative
sensitivity of differing object areas {tiny, small, medium, large} with respect
to the compression level. We show that tiny and small objects are more
sensitive to compression than medium and large objects. Overall, Cascade R-CNN
attains the maximal mAP across most of the object area categories.
Related papers
- Implicit Neural Image Field for Biological Microscopy Image Compression [37.0218688308699]
We propose an adaptive compression workflow based on Implicit Neural Representation (INR)
This approach permits application-specific compression objectives, capable of compressing images of any shape and arbitrary pixel-wise decompression.
We demonstrated on a wide range of microscopy images that our workflow not only achieved high, controllable compression ratios but also preserved detailed information critical for downstream analysis.
arXiv Detail & Related papers (2024-05-29T11:51:33Z) - Deep learning based Image Compression for Microscopy Images: An
Empirical Study [3.915183869199319]
This study analyzes classic and deep learning based image compression methods, and their impact on deep learning based image processing models.
To compress images in such a wanted way, multiple classical lossy image compression techniques are compared to several AI-based compression models.
We found that AI-based compression techniques largely outperform the classic ones and will minimally affect the downstream label-free task in 2D cases.
arXiv Detail & Related papers (2023-11-02T16:00:32Z) - Are Visual Recognition Models Robust to Image Compression? [23.280147529096908]
We analyze the impact of image compression on visual recognition tasks.
We consider a wide range of compression levels, ranging from 0.1 to 2 bits-per-pixel (bpp)
We find that for all three tasks, the recognition ability is significantly impacted when using strong compression.
arXiv Detail & Related papers (2023-04-10T11:30:11Z) - Learned Lossless Compression for JPEG via Frequency-Domain Prediction [50.20577108662153]
We propose a novel framework for learned lossless compression of JPEG images.
To enable learning in the frequency domain, DCT coefficients are partitioned into groups to utilize implicit local redundancy.
An autoencoder-like architecture is designed based on the weight-shared blocks to realize entropy modeling of grouped DCT coefficients.
arXiv Detail & Related papers (2023-03-05T13:15:28Z) - Crowd Counting on Heavily Compressed Images with Curriculum Pre-Training [90.76576712433595]
Applying lossy compression on images processed by deep neural networks can lead to significant accuracy degradation.
Inspired by the curriculum learning paradigm, we present a novel training approach called curriculum pre-training (CPT) for crowd counting on compressed images.
arXiv Detail & Related papers (2022-08-15T08:43:21Z) - Towards Compact CNNs via Collaborative Compression [166.86915086497433]
We propose a Collaborative Compression scheme, which joints channel pruning and tensor decomposition to compress CNN models.
We achieve 52.9% FLOPs reduction by removing 48.4% parameters on ResNet-50 with only a Top-1 accuracy drop of 0.56% on ImageNet 2012.
arXiv Detail & Related papers (2021-05-24T12:07:38Z) - Learning Scalable $\ell_\infty$-constrained Near-lossless Image
Compression via Joint Lossy Image and Residual Compression [118.89112502350177]
We propose a novel framework for learning $ell_infty$-constrained near-lossless image compression.
We derive the probability model of the quantized residual by quantizing the learned probability model of the original residual.
arXiv Detail & Related papers (2021-03-31T11:53:36Z) - Towards Robust Data Hiding Against (JPEG) Compression: A
Pseudo-Differentiable Deep Learning Approach [78.05383266222285]
It is still an open challenge to achieve the goal of data hiding that can be against these compressions.
Deep learning has shown large success in data hiding, while non-differentiability of JPEG makes it challenging to train a deep pipeline for improving robustness against lossy compression.
In this work, we propose a simple yet effective approach to address all the above limitations at once.
arXiv Detail & Related papers (2020-12-30T12:30:09Z) - Analyzing and Mitigating JPEG Compression Defects in Deep Learning [69.04777875711646]
We present a unified study of the effects of JPEG compression on a range of common tasks and datasets.
We show that there is a significant penalty on common performance metrics for high compression.
arXiv Detail & Related papers (2020-11-17T20:32:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.