Data Efficient Visual Place Recognition Using Extremely JPEG-Compressed
Images
- URL: http://arxiv.org/abs/2209.08343v1
- Date: Sat, 17 Sep 2022 14:46:28 GMT
- Title: Data Efficient Visual Place Recognition Using Extremely JPEG-Compressed
Images
- Authors: Mihnea-Alexandru Tomita, Bruno Ferrarini, Michael Milford, Klaus
McDonald-Maier, Shoaib Ehsan
- Abstract summary: This paper studies the effects of JPEG compression on the performance of Visual Place Recognition techniques.
We show that by introducing compression, the VPR performance is drastically reduced, especially in the higher spectrum of compression.
We present a fine-tuned CNN which is optimized for JPEG compressed data and show that it performs more consistently with the image transformations detected in extremely compressed JPEG images.
- Score: 17.847661026367767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual Place Recognition (VPR) is the ability of a robotic platform to
correctly interpret visual stimuli from its on-board cameras in order to
determine whether it is currently located in a previously visited place,
despite different viewpoint, illumination and appearance changes. JPEG is a
widely used image compression standard that is capable of significantly
reducing the size of an image at the cost of image clarity. For applications
where several robotic platforms are simultaneously deployed, the visual data
gathered must be transmitted remotely between each robot. Hence, JPEG
compression can be employed to drastically reduce the amount of data
transmitted over a communication channel, as working with limited bandwidth for
VPR can be proven to be a challenging task. However, the effects of JPEG
compression on the performance of current VPR techniques have not been
previously studied. For this reason, this paper presents an in-depth study of
JPEG compression in VPR related scenarios. We use a selection of
well-established VPR techniques on 8 datasets with various amounts of
compression applied. We show that by introducing compression, the VPR
performance is drastically reduced, especially in the higher spectrum of
compression. To overcome the negative effects of JPEG compression on the VPR
performance, we present a fine-tuned CNN which is optimized for JPEG compressed
data and show that it performs more consistently with the image transformations
detected in extremely compressed JPEG images.
Related papers
- Learned Image Compression for HE-stained Histopathological Images via Stain Deconvolution [33.69980388844034]
In this paper, we show that the commonly used JPEG algorithm is not best suited for further compression.
We propose Stain Quantized Latent Compression, a novel DL based histopathology data compression approach.
We show that our approach yields superior performance in a classification downstream task, compared to traditional approaches like JPEG.
arXiv Detail & Related papers (2024-06-18T13:47:17Z) - PatchSVD: A Non-uniform SVD-based Image Compression Algorithm [20.856903918492154]
We propose a novel region-based lossy image compression technique, called PatchSVD, based on the Singular Value Decomposition (SVD) algorithm.
We show through experiments that PatchSVD outperforms SVD-based image compression with respect to three popular image compression metrics.
arXiv Detail & Related papers (2024-06-07T17:57:40Z) - Are Visual Recognition Models Robust to Image Compression? [23.280147529096908]
We analyze the impact of image compression on visual recognition tasks.
We consider a wide range of compression levels, ranging from 0.1 to 2 bits-per-pixel (bpp)
We find that for all three tasks, the recognition ability is significantly impacted when using strong compression.
arXiv Detail & Related papers (2023-04-10T11:30:11Z) - Learned Lossless Compression for JPEG via Frequency-Domain Prediction [50.20577108662153]
We propose a novel framework for learned lossless compression of JPEG images.
To enable learning in the frequency domain, DCT coefficients are partitioned into groups to utilize implicit local redundancy.
An autoencoder-like architecture is designed based on the weight-shared blocks to realize entropy modeling of grouped DCT coefficients.
arXiv Detail & Related papers (2023-03-05T13:15:28Z) - Data-Efficient Sequence-Based Visual Place Recognition with Highly
Compressed JPEG Images [17.847661026367767]
Visual Place Recognition (VPR) is a fundamental task that allows a robotic platform to successfully localise itself in the environment.
JPEG is an image compression standard that can employ high compression ratios to facilitate lower data transmission for VPR applications.
When applying high levels of JPEG compression, both the image clarity and size are drastically reduced.
arXiv Detail & Related papers (2023-02-26T13:13:51Z) - Crowd Counting on Heavily Compressed Images with Curriculum Pre-Training [90.76576712433595]
Applying lossy compression on images processed by deep neural networks can lead to significant accuracy degradation.
Inspired by the curriculum learning paradigm, we present a novel training approach called curriculum pre-training (CPT) for crowd counting on compressed images.
arXiv Detail & Related papers (2022-08-15T08:43:21Z) - Towards Robust Data Hiding Against (JPEG) Compression: A
Pseudo-Differentiable Deep Learning Approach [78.05383266222285]
It is still an open challenge to achieve the goal of data hiding that can be against these compressions.
Deep learning has shown large success in data hiding, while non-differentiability of JPEG makes it challenging to train a deep pipeline for improving robustness against lossy compression.
In this work, we propose a simple yet effective approach to address all the above limitations at once.
arXiv Detail & Related papers (2020-12-30T12:30:09Z) - Analyzing and Mitigating JPEG Compression Defects in Deep Learning [69.04777875711646]
We present a unified study of the effects of JPEG compression on a range of common tasks and datasets.
We show that there is a significant penalty on common performance metrics for high compression.
arXiv Detail & Related papers (2020-11-17T20:32:57Z) - Learning to Improve Image Compression without Changing the Standard
Decoder [100.32492297717056]
We propose learning to improve the encoding performance with the standard decoder.
Specifically, a frequency-domain pre-editing method is proposed to optimize the distribution of DCT coefficients.
We do not modify the JPEG decoder and therefore our approach is applicable when viewing images with the widely used standard JPEG decoder.
arXiv Detail & Related papers (2020-09-27T19:24:42Z) - Discernible Image Compression [124.08063151879173]
This paper aims to produce compressed images by pursuing both appearance and perceptual consistency.
Based on the encoder-decoder framework, we propose using a pre-trained CNN to extract features of the original and compressed images.
Experiments on benchmarks demonstrate that images compressed by using the proposed method can also be well recognized by subsequent visual recognition and detection models.
arXiv Detail & Related papers (2020-02-17T07:35:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.