Deep Dict: Deep Learning-based Lossy Time Series Compressor for IoT Data
- URL: http://arxiv.org/abs/2401.10396v1
- Date: Thu, 18 Jan 2024 22:10:21 GMT
- Title: Deep Dict: Deep Learning-based Lossy Time Series Compressor for IoT Data
- Authors: Jinxin Liu, Petar Djukic, Michel Kulhandjian, Burak Kantarci
- Abstract summary: Deep Dict is a lossy time series compressor designed to achieve a high compression ratio while maintaining decompression error within a predefined range.
BTAE extracts Bernoulli representations from time series data, reducing the size of the representations compared to conventional autoencoders.
In order to address the limitations of common regression losses such as L1/L2, we introduce a novel loss function called quantized entropy loss (QEL)
- Score: 15.97162100346596
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We propose Deep Dict, a deep learning-based lossy time series compressor
designed to achieve a high compression ratio while maintaining decompression
error within a predefined range. Deep Dict incorporates two essential
components: the Bernoulli transformer autoencoder (BTAE) and a distortion
constraint. BTAE extracts Bernoulli representations from time series data,
reducing the size of the representations compared to conventional autoencoders.
The distortion constraint limits the prediction error of BTAE to the desired
range. Moreover, in order to address the limitations of common regression
losses such as L1/L2, we introduce a novel loss function called quantized
entropy loss (QEL). QEL takes into account the specific characteristics of the
problem, enhancing robustness to outliers and alleviating optimization
challenges. Our evaluation of Deep Dict across ten diverse time series datasets
from various domains reveals that Deep Dict outperforms state-of-the-art lossy
compressors in terms of compression ratio by a significant margin by up to
53.66%.
Related papers
- CAMEO: Autocorrelation-Preserving Line Simplification for Lossy Time Series Compression [7.938342455750219]
We propose a new lossy compression method that provides guarantees on the autocorrelation and partial-autocorrelation functions of a time series.
Our method improves compression ratios by 2x on average and up to 54x on selected datasets.
arXiv Detail & Related papers (2025-01-24T11:59:51Z) - CALLIC: Content Adaptive Learning for Lossless Image Compression [64.47244912937204]
CALLIC sets a new state-of-the-art (SOTA) for learned lossless image compression.
We propose a content-aware autoregressive self-attention mechanism by leveraging convolutional gating operations.
During encoding, we decompose pre-trained layers, including depth-wise convolutions, using low-rank matrices and then adapt the incremental weights on testing image by Rate-guided Progressive Fine-Tuning (RPFT)
RPFT fine-tunes with gradually increasing patches that are sorted in descending order by estimated entropy, optimizing learning process and reducing adaptation time.
arXiv Detail & Related papers (2024-12-23T10:41:18Z) - Learned Compression of Nonlinear Time Series With Random Access [2.564905016909138]
Time series play a crucial role in many fields, including finance, healthcare, industry, and environmental monitoring.
We introduce NeaTS, a randomly-accessible compression scheme that approximates the time series with a sequence of nonlinear functions.
Our experiments show that NeaTS improves the compression ratio of the state-of-the-art lossy compressors by up to 14%.
arXiv Detail & Related papers (2024-12-20T10:30:06Z) - Variable Rate Neural Compression for Sparse Detector Data [9.331686712558144]
We propose a novel approach for TPC data compression via key-point identification facilitated by sparse convolution.
BCAE-VS achieves a $75%$ improvement in reconstruction accuracy with a $10%$ increase in compression ratio over the previous state-of-the-art model.
arXiv Detail & Related papers (2024-11-18T17:15:35Z) - Unlocking Data-free Low-bit Quantization with Matrix Decomposition for KV Cache Compression [87.5604418100301]
Key-value( KV) caching is an important technique to accelerate the inference of large language models.
Existing methods often compromise precision or require extra data for calibration.
We introduce textbfDecoQuant, a novel data-free low-bit quantization technique based on tensor decomposition methods.
arXiv Detail & Related papers (2024-05-21T08:35:10Z) - Convolutional variational autoencoders for secure lossy image compression in remote sensing [47.75904906342974]
This study investigates image compression based on convolutional variational autoencoders (CVAE)
CVAEs have been demonstrated to outperform conventional compression methods such as JPEG2000 by a substantial margin on compression benchmark datasets.
arXiv Detail & Related papers (2024-04-03T15:17:29Z) - Compression of Structured Data with Autoencoders: Provable Benefit of
Nonlinearities and Depth [83.15263499262824]
We prove that gradient descent converges to a solution that completely disregards the sparse structure of the input.
We show how to improve upon Gaussian performance for the compression of sparse data by adding a denoising function to a shallow architecture.
We validate our findings on image datasets, such as CIFAR-10 and MNIST.
arXiv Detail & Related papers (2024-02-07T16:32:29Z) - Inshrinkerator: Compressing Deep Learning Training Checkpoints via Dynamic Quantization [5.648270790530862]
State-of-the-art approaches involve lossy model compression mechanisms, which induce a tradeoff between the resulting model quality (accuracy) and compression ratio.
We make a key enabling observation that the sensitivity of model weights to compression varies during training, and different weights benefit from different quantization levels.
We propose a non-uniform quantization scheme that leverages this variation, an efficient search mechanism that dynamically finds the best quantization configurations, and a quantization-aware delta compression mechanism that rearranges weights to minimize checkpoint differences.
arXiv Detail & Related papers (2023-06-20T18:00:31Z) - Unrolled Compressed Blind-Deconvolution [77.88847247301682]
sparse multichannel blind deconvolution (S-MBD) arises frequently in many engineering applications such as radar/sonar/ultrasound imaging.
We propose a compression method that enables blind recovery from much fewer measurements with respect to the full received signal in time.
arXiv Detail & Related papers (2022-09-28T15:16:58Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.