Image Storage on Synthetic DNA Using Autoencoders
- URL: http://arxiv.org/abs/2203.09981v1
- Date: Fri, 18 Mar 2022 14:17:48 GMT
- Title: Image Storage on Synthetic DNA Using Autoencoders
- Authors: Xavier Pic and Marc Antonini
- Abstract summary: This paper presents some results on lossy image compression methods based on convolutional autoencoders adapted to DNA data storage.
The model architectures presented here have been designed to efficiently compress images, encode them into a quaternary code, and finally store them into synthetic DNA molecules.
- Score: 6.096779295981377
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Over the past years, the ever-growing trend on data storage demand, more
specifically for "cold" data (rarely accessed data), has motivated research for
alternative systems of data storage. Because of its biochemical
characteristics, synthetic DNA molecules are now considered as serious
candidates for this new kind of storage. This paper presents some results on
lossy image compression methods based on convolutional autoencoders adapted to
DNA data storage.
The model architectures presented here have been designed to efficiently
compress images, encode them into a quaternary code, and finally store them
into synthetic DNA molecules. This work also aims at making the compression
models better fit the problematics that we encounter when storing data into
DNA, namely the fact that the DNA writing, storing and reading methods are
error prone processes. The main take away of this kind of compressive
autoencoder is our quantization and the robustness to substitution errors
thanks to the noise model that we use during training.
Related papers
- A Bird-Eye view on DNA Storage Simulators [0.0]
This paper aims to review some of the software that performs DNA storage simulations in different domains.
We present 3 different softwares on the basis of domain, implementation techniques, and customer/commercial usability.
arXiv Detail & Related papers (2024-04-07T08:46:42Z) - Neural-based Compression Scheme for Solar Image Data [8.374518151411612]
We propose a neural network-based lossy compression method to be used in NASA's data-intensive imagery missions.
In this work, we propose an adversarially trained neural network, equipped with local and non-local attention modules to capture both the local and global structure of the image.
As a proof of concept for use of this algorithm in SDO data analysis, we have performed coronal hole (CH) detection using our compressed images.
arXiv Detail & Related papers (2023-11-06T04:13:58Z) - Implicit Neural Multiple Description for DNA-based data storage [6.423239719448169]
DNA exhibits remarkable potential as a data storage solution due to its impressive storage density and long-term stability.
However, developing this novel medium comes with its own set of challenges, particularly in addressing errors arising from storage and biological manipulations.
We have pioneered a novel compression scheme and a cutting-edge Multiple Description Coding (MDC) technique utilizing neural networks for DNA data storage.
arXiv Detail & Related papers (2023-09-13T13:42:52Z) - DNA data storage, sequencing data-carrying DNA [2.4493299476776778]
We study accuracy trade-offs between deep model size and error correcting codes.
We show that, starting with a model size of 107MB, the reduced accuracy from model compression can be compensated by using simple error correcting codes.
arXiv Detail & Related papers (2022-05-11T13:31:57Z) - Memory Replay with Data Compression for Continual Learning [80.95444077825852]
We propose memory replay with data compression to reduce the storage cost of old training samples.
We extensively validate this across several benchmarks of class-incremental learning and in a realistic scenario of object detection for autonomous driving.
arXiv Detail & Related papers (2022-02-14T10:26:23Z) - COIN++: Data Agnostic Neural Compression [55.27113889737545]
COIN++ is a neural compression framework that seamlessly handles a wide range of data modalities.
We demonstrate the effectiveness of our method by compressing various data modalities.
arXiv Detail & Related papers (2022-01-30T20:12:04Z) - Single-Read Reconstruction for DNA Data Storage Using Transformers [0.0]
We propose a novel approach for single-read reconstruction using an encoder-decoder Transformer architecture for DNA based data storage.
Our model achieves lower error rates when reconstructing the original data from a single read of each DNA strand.
This is the first demonstration of using deep learning models for single-read reconstruction in DNA based storage.
arXiv Detail & Related papers (2021-09-12T10:01:59Z) - Deep DNA Storage: Scalable and Robust DNA Storage via Coding Theory and
Deep Learning [49.3231734733112]
We show a modular and holistic approach that combines Deep Neural Networks (DNN) trained on simulated data, Product (TP) based Error-Correcting Codes (ECC) and a safety margin into a single coherent pipeline.
Our work improves upon the current leading solutions by up to x3200 increase in speed, 40% improvement in accuracy, and offers a code rate of 1.6 bits per base in a high noise regime.
arXiv Detail & Related papers (2021-08-31T18:21:20Z) - Towards Robust Data Hiding Against (JPEG) Compression: A
Pseudo-Differentiable Deep Learning Approach [78.05383266222285]
It is still an open challenge to achieve the goal of data hiding that can be against these compressions.
Deep learning has shown large success in data hiding, while non-differentiability of JPEG makes it challenging to train a deep pipeline for improving robustness against lossy compression.
In this work, we propose a simple yet effective approach to address all the above limitations at once.
arXiv Detail & Related papers (2020-12-30T12:30:09Z) - Analyzing and Mitigating JPEG Compression Defects in Deep Learning [69.04777875711646]
We present a unified study of the effects of JPEG compression on a range of common tasks and datasets.
We show that there is a significant penalty on common performance metrics for high compression.
arXiv Detail & Related papers (2020-11-17T20:32:57Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.