Interpreting Training Aspects of Deep-Learned Error-Correcting Codes
- URL: http://arxiv.org/abs/2305.04347v1
- Date: Sun, 7 May 2023 17:53:31 GMT
- Title: Interpreting Training Aspects of Deep-Learned Error-Correcting Codes
- Authors: N. Devroye, A. Mulgund, R. Shekhar, Gy. Tur\'an, M. \v{Z}efran, Y.
Zhou
- Abstract summary: We look at developing tools for interpreting the training process for deep-learned error-correcting codes.
All tools are demonstrated on TurboAE, but are applicable to other deep-learned forward error correcting codes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As new deep-learned error-correcting codes continue to be introduced, it is
important to develop tools to interpret the designed codes and understand the
training process. Prior work focusing on the deep-learned TurboAE has both
interpreted the learned encoders post-hoc by mapping these onto nearby
``interpretable'' encoders, and experimentally evaluated the performance of
these interpretable encoders with various decoders. Here we look at developing
tools for interpreting the training process for deep-learned error-correcting
codes, focusing on: 1) using the Goldreich-Levin algorithm to quickly interpret
the learned encoder; 2) using Fourier coefficients as a tool for understanding
the training dynamics and the loss landscape; 3) reformulating the training
loss, the binary cross entropy, by relating it to encoder and decoder
parameters, and the bit error rate (BER); 4) using these insights to formulate
and study a new training procedure. All tools are demonstrated on TurboAE, but
are applicable to other deep-learned forward error correcting codes (without
feedback).
Related papers
- Collective Bit Flipping-Based Decoding of Quantum LDPC Codes [0.6554326244334866]
We improve both the error correction performance and decoding latency of variable degree-3 (dv-3) QLDPC codes under iterative decoding.
Our decoding scheme is based on applying a modified version of bit flipping (BF) decoding, namely two-bit bit flipping (TBF) decoding.
arXiv Detail & Related papers (2024-06-24T18:51:48Z) - Learning Linear Block Error Correction Codes [62.25533750469467]
We propose for the first time a unified encoder-decoder training of binary linear block codes.
We also propose a novel Transformer model in which the self-attention masking is performed in a differentiable fashion for the efficient backpropagation of the code gradient.
arXiv Detail & Related papers (2024-05-07T06:47:12Z) - Graph Neural Networks for Enhanced Decoding of Quantum LDPC Codes [6.175503577352742]
We propose a differentiable iterative decoder for quantum low-density parity-check (LDPC) codes.
The proposed algorithm is composed of classical belief propagation (BP) decoding stages and intermediate graph neural network (GNN) layers.
arXiv Detail & Related papers (2023-10-26T19:56:25Z) - A Scalable Graph Neural Network Decoder for Short Block Codes [49.25571364253986]
We propose a novel decoding algorithm for short block codes based on an edge-weighted graph neural network (EW-GNN)
The EW-GNN decoder operates on the Tanner graph with an iterative message-passing structure.
We show that the EW-GNN decoder outperforms the BP and deep-learning-based BP methods in terms of the decoding error rate.
arXiv Detail & Related papers (2022-11-13T17:13:12Z) - Graph Neural Networks for Channel Decoding [71.15576353630667]
We showcase competitive decoding performance for various coding schemes, such as low-density parity-check (LDPC) and BCH codes.
The idea is to let a neural network (NN) learn a generalized message passing algorithm over a given graph.
We benchmark our proposed decoder against state-of-the-art in conventional channel decoding as well as against recent deep learning-based results.
arXiv Detail & Related papers (2022-07-29T15:29:18Z) - Pre-Training Transformer Decoder for End-to-End ASR Model with Unpaired
Speech Data [145.95460945321253]
We introduce two pre-training tasks for the encoder-decoder network using acoustic units, i.e., pseudo codes.
The proposed Speech2C can relatively reduce the word error rate (WER) by 19.2% over the method without decoder pre-training.
arXiv Detail & Related papers (2022-03-31T15:33:56Z) - A manifold learning perspective on representation learning: Learning
decoder and representations without an encoder [0.0]
Autoencoders are commonly used in representation learning.
Inspired by manifold learning, we show that the decoder can be trained on its own by learning the representations of the training samples.
Our approach of training the decoder alone facilitates representation learning even on small data sets.
arXiv Detail & Related papers (2021-08-31T15:08:50Z) - Deep Learning to Ternary Hash Codes by Continuation [8.920717493647121]
We propose to jointly learn the features with the codes by appending a smoothed function to the networks.
Experiments show that the generated codes indeed could achieve higher retrieval accuracy.
arXiv Detail & Related papers (2021-07-16T16:02:08Z) - Scheduled Sampling in Vision-Language Pretraining with Decoupled
Encoder-Decoder Network [99.03895740754402]
We propose a two-stream decoupled design of encoder-decoder structure, in which two decoupled cross-modal encoder and decoder are involved.
As an alternative, we propose a primary scheduled sampling strategy that mitigates such discrepancy via pretraining encoder-decoder in a two-pass manner.
arXiv Detail & Related papers (2021-01-27T17:36:57Z) - On the Sub-Layer Functionalities of Transformer Decoder [74.83087937309266]
We study how Transformer-based decoders leverage information from the source and target languages.
Based on these insights, we demonstrate that the residual feed-forward module in each Transformer decoder layer can be dropped with minimal loss of performance.
arXiv Detail & Related papers (2020-10-06T11:50:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.