Less is More: Accelerating Faster Neural Networks Straight from JPEG
- URL: http://arxiv.org/abs/2104.00185v1
- Date: Thu, 1 Apr 2021 01:21:24 GMT
- Title: Less is More: Accelerating Faster Neural Networks Straight from JPEG
- Authors: Samuel Felipe dos Santos and Jurandy Almeida
- Abstract summary: We show how to speed up convolutional neural networks for processing JPEG compressed data.
We exploit learning strategies to reduce the computational complexity by taking full advantage of DCT inputs.
Results show that learning how to combine all DCT inputs in a data-driven fashion is better than discarding them by hand.
- Score: 1.9214041945441434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most image data available are often stored in a compressed format, from which
JPEG is the most widespread. To feed this data on a convolutional neural
network (CNN), a preliminary decoding process is required to obtain RGB pixels,
demanding a high computational load and memory usage. For this reason, the
design of CNNs for processing JPEG compressed data has gained attention in
recent years. In most existing works, typical CNN architectures are adapted to
facilitate the learning with the DCT coefficients rather than RGB pixels.
Although they are effective, their architectural changes either raise the
computational costs or neglect relevant information from DCT inputs. In this
paper, we examine different ways of speeding up CNNs designed for DCT inputs,
exploiting learning strategies to reduce the computational complexity by taking
full advantage of DCT inputs. Our experiments were conducted on the ImageNet
dataset. Results show that learning how to combine all DCT inputs in a
data-driven fashion is better than discarding them by hand, and its combination
with a reduction of layers has proven to be effective for reducing the
computational costs while retaining accuracy.
Related papers
- CNNs for JPEGs: A Study in Computational Cost [45.74830585715129]
Convolutional neural networks (CNNs) have achieved astonishing advances over the past decade.
CNNs are capable of learning robust representations of the data directly from the RGB pixels.
Deep learning methods capable of learning directly from the compressed domain have been gaining attention in recent years.
arXiv Detail & Related papers (2023-09-20T15:49:38Z) - Dataset Quantization [72.61936019738076]
We present dataset quantization (DQ), a new framework to compress large-scale datasets into small subsets.
DQ is the first method that can successfully distill large-scale datasets such as ImageNet-1k with a state-of-the-art compression ratio.
arXiv Detail & Related papers (2023-08-21T07:24:29Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Convolutional Neural Network (CNN) to reduce construction loss in JPEG
compression caused by Discrete Fourier Transform (DFT) [0.0]
Convolutional Neural Networks (CNN) have received more attention than most other types of deep neural networks.
In this work, an effective image compression method is purposed using autoencoders.
arXiv Detail & Related papers (2022-08-26T12:46:16Z) - COIN++: Data Agnostic Neural Compression [55.27113889737545]
COIN++ is a neural compression framework that seamlessly handles a wide range of data modalities.
We demonstrate the effectiveness of our method by compressing various data modalities.
arXiv Detail & Related papers (2022-01-30T20:12:04Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - Learning JPEG Compression Artifacts for Image Manipulation Detection and
Localization [26.36646590957593]
It is essential to distinguish between authentic and tampered regions by analyzing intrinsic statistics in an image.
We propose a convolutional neural network (CNN) that uses discrete cosine transform (DCT) coefficients, where compression artifacts remain, to localize image manipulation.
We show how to design and train a neural network that can learn the distribution of DCT coefficients.
It significantly outperforms traditional and deep neural network-based methods in detecting and localizing tampered regions.
arXiv Detail & Related papers (2021-08-30T01:21:07Z) - CNNs for JPEGs: A Study in Computational Cost [49.97673761305336]
Convolutional neural networks (CNNs) have achieved astonishing advances over the past decade.
CNNs are capable of learning robust representations of the data directly from the RGB pixels.
Deep learning methods capable of learning directly from the compressed domain have been gaining attention in recent years.
arXiv Detail & Related papers (2020-12-26T15:00:10Z) - DCT-SNN: Using DCT to Distribute Spatial Information over Time for
Learning Low-Latency Spiking Neural Networks [7.876001630578417]
Spiking Neural Networks (SNNs) offer a promising alternative to traditional deep learning frameworks.
SNNs suffer from high inference latency which is a major bottleneck to their deployment.
We propose a scalable time-based encoding scheme that utilizes the Discrete Cosine Transform (DCT) to reduce the number of timesteps required for inference.
arXiv Detail & Related papers (2020-10-05T05:55:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.