Efficient Data Compression for 3D Sparse TPC via Bicephalous
Convolutional Autoencoder
- URL: http://arxiv.org/abs/2111.05423v1
- Date: Tue, 9 Nov 2021 21:26:37 GMT
- Title: Efficient Data Compression for 3D Sparse TPC via Bicephalous
Convolutional Autoencoder
- Authors: Yi Huang, Yihui Ren, Shinjae Yoo, Jin Huang
- Abstract summary: This work introduces a dual-head autoencoder to resolve sparsity and regression simultaneously, called textitBicephalous Convolutional AutoEncoder (BCAE)
It shows advantages both in compression fidelity and ratio compared to traditional data compression methods, such as MGARD, SZ, and ZFP.
- Score: 8.759778406741276
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-time data collection and analysis in large experimental facilities
present a great challenge across multiple domains, including high energy
physics, nuclear physics, and cosmology. To address this, machine learning
(ML)-based methods for real-time data compression have drawn significant
attention. However, unlike natural image data, such as CIFAR and ImageNet that
are relatively small-sized and continuous, scientific data often come in as
three-dimensional data volumes at high rates with high sparsity (many zeros)
and non-Gaussian value distribution. This makes direct application of popular
ML compression methods, as well as conventional data compression methods,
suboptimal. To address these obstacles, this work introduces a dual-head
autoencoder to resolve sparsity and regression simultaneously, called
\textit{Bicephalous Convolutional AutoEncoder} (BCAE). This method shows
advantages both in compression fidelity and ratio compared to traditional data
compression methods, such as MGARD, SZ, and ZFP. To achieve similar fidelity,
the best performer among the traditional methods can reach only half the
compression ratio of BCAE. Moreover, a thorough ablation study of the BCAE
method shows that a dedicated segmentation decoder improves the reconstruction.
Related papers
- Attention Based Machine Learning Methods for Data Reduction with Guaranteed Error Bounds [11.494915987840876]
Scientific applications in fields such as high energy physics generate vast amounts of data at high velocities.
To address this challenge, data compression or reduction techniques are crucial.
We propose an attention-based compression method utilizing a blockwise compression setup.
arXiv Detail & Related papers (2024-09-09T06:35:24Z) - UniCompress: Enhancing Multi-Data Medical Image Compression with Knowledge Distillation [59.3877309501938]
Implicit Neural Representation (INR) networks have shown remarkable versatility due to their flexible compression ratios.
We introduce a codebook containing frequency domain information as a prior input to the INR network.
This enhances the representational power of INR and provides distinctive conditioning for different image blocks.
arXiv Detail & Related papers (2024-05-27T05:52:13Z) - Sparse $L^1$-Autoencoders for Scientific Data Compression [0.0]
We introduce effective data compression methods by developing autoencoders using high dimensional latent spaces that are $L1$-regularized.
We show how these information-rich latent spaces can be used to mitigate blurring and other artifacts to obtain highly effective data compression methods for scientific data.
arXiv Detail & Related papers (2024-05-23T07:48:00Z) - Convolutional variational autoencoders for secure lossy image compression in remote sensing [47.75904906342974]
This study investigates image compression based on convolutional variational autoencoders (CVAE)
CVAEs have been demonstrated to outperform conventional compression methods such as JPEG2000 by a substantial margin on compression benchmark datasets.
arXiv Detail & Related papers (2024-04-03T15:17:29Z) - Neural-based Compression Scheme for Solar Image Data [8.374518151411612]
We propose a neural network-based lossy compression method to be used in NASA's data-intensive imagery missions.
In this work, we propose an adversarially trained neural network, equipped with local and non-local attention modules to capture both the local and global structure of the image.
As a proof of concept for use of this algorithm in SDO data analysis, we have performed coronal hole (CH) detection using our compressed images.
arXiv Detail & Related papers (2023-11-06T04:13:58Z) - Fast 2D Bicephalous Convolutional Autoencoder for Compressing 3D Time
Projection Chamber Data [11.186303973102532]
This work introduces two BCAE variants: BCAE++ and BCAE-2D.
BCAE++ achieves a 15% better compression ratio and a 77% better reconstruction accuracy measured in mean absolute error compared with BCAE.
In addition, we demonstrate an unbalanced autoencoder with a larger decoder can improve reconstruction accuracy without significantly sacrificing throughput.
arXiv Detail & Related papers (2023-10-23T15:23:32Z) - Quick Dense Retrievers Consume KALE: Post Training Kullback Leibler
Alignment of Embeddings for Asymmetrical dual encoders [89.29256833403169]
We introduce Kullback Leibler Alignment of Embeddings (KALE), an efficient and accurate method for increasing the inference efficiency of dense retrieval methods.
KALE extends traditional Knowledge Distillation after bi-encoder training, allowing for effective query encoder compression without full retraining or index generation.
Using KALE and asymmetric training, we can generate models which exceed the performance of DistilBERT despite having 3x faster inference.
arXiv Detail & Related papers (2023-03-31T15:44:13Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Exploring Autoencoder-based Error-bounded Compression for Scientific
Data [14.724393511470225]
We develop an error-bounded autoencoder-based framework in terms of the SZ model.
We optimize the compression quality for the main stages in our designed AE-based error-bounded compression framework.
arXiv Detail & Related papers (2021-05-25T07:53:32Z) - PowerGossip: Practical Low-Rank Communication Compression in
Decentralized Deep Learning [62.440827696638664]
We introduce a simple algorithm that directly compresses the model differences between neighboring workers.
Inspired by the PowerSGD for centralized deep learning, this algorithm uses power steps to maximize the information transferred per bit.
arXiv Detail & Related papers (2020-08-04T09:14:52Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.