Autoregressive Belief Propagation for Decoding Block Codes
- URL: http://arxiv.org/abs/2103.11780v1
- Date: Sat, 23 Jan 2021 17:14:55 GMT
- Title: Autoregressive Belief Propagation for Decoding Block Codes
- Authors: Eliya Nachmani, Lior Wolf
- Abstract summary: We revisit recent methods that employ graph neural networks for decoding error correcting codes.
Our method violates the symmetry conditions that enable the other methods to train exclusively with the zero-word.
Despite not having the luxury of training on a single word, and the inability to train on more than a small fraction of the relevant sample space, we demonstrate effective training.
- Score: 113.38181979662288
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We revisit recent methods that employ graph neural networks for decoding
error correcting codes and employ messages that are computed in an
autoregressive manner. The outgoing messages of the variable nodes are
conditioned not only on the incoming messages, but also on an estimation of the
SNR and on the inferred codeword and on two downstream computations: (i) an
extended vector of parity check outcomes, (ii) the mismatch between the
inferred codeword and the re-encoding of the information bits of this codeword.
Unlike most learned methods in the field, our method violates the symmetry
conditions that enable the other methods to train exclusively with the
zero-word. Despite not having the luxury of training on a single word, and the
inability to train on more than a small fraction of the relevant sample space,
we demonstrate effective training. The new method obtains a bit error rate that
outperforms the latest methods by a sizable margin.
Related papers
- A blindness property of the Min-Sum decoding for the toric code [3.543432625843538]
Kitaev's toric code is one of the most prominent models for fault-tolerant quantum computation.
Recent efforts have been devoted to improving the error correction performance of the toric code under message-passing decoding.
arXiv Detail & Related papers (2024-06-21T08:28:31Z) - Bit Cipher -- A Simple yet Powerful Word Representation System that
Integrates Efficiently with Language Models [4.807347156077897]
Bit-cipher is a word representation system that eliminates the need of backpropagation and hyper-efficient dimensionality reduction techniques.
We perform probing experiments on part-of-speech (POS) tagging and named entity recognition (NER) to assess bit-cipher's competitiveness with classic embeddings.
By replacing embedding layers with cipher embeddings, our experiments illustrate the notable efficiency of cipher in accelerating the training process and attaining better optima.
arXiv Detail & Related papers (2023-11-18T08:47:35Z) - GEC-DePenD: Non-Autoregressive Grammatical Error Correction with
Decoupled Permutation and Decoding [52.14832976759585]
Grammatical error correction (GEC) is an important NLP task that is usually solved with autoregressive sequence-to-sequence models.
We propose a novel non-autoregressive approach to GEC that decouples the architecture into a permutation network.
We show that the resulting network improves over previously known non-autoregressive methods for GEC.
arXiv Detail & Related papers (2023-11-14T14:24:36Z) - Downstream Task-Oriented Neural Tokenizer Optimization with Vocabulary
Restriction as Post Processing [4.781986758380065]
This paper proposes a method to optimize tokenization for the performance improvement of already trained downstream models.
Our method generates tokenization results attaining lower loss values of a given downstream model on the training data for restricting vocabularies and trains a tokenizer reproducing the tokenization results.
arXiv Detail & Related papers (2023-04-21T08:29:14Z) - Provably Convergent Subgraph-wise Sampling for Fast GNN Training [122.68566970275683]
We propose a novel subgraph-wise sampling method with a convergence guarantee, namely Local Message Compensation (LMC)
LMC retrieves the discarded messages in backward passes based on a message passing formulation of backward passes.
Experiments on large-scale benchmarks demonstrate that LMC is significantly faster than state-of-the-art subgraph-wise sampling methods.
arXiv Detail & Related papers (2023-03-17T05:16:49Z) - Soft-Labeled Contrastive Pre-training for Function-level Code
Representation [127.71430696347174]
We present textbfSCodeR, a textbfSoft-labeled contrastive pre-training framework with two positive sample construction methods.
Considering the relevance between codes in a large-scale code corpus, the soft-labeled contrastive pre-training can obtain fine-grained soft-labels.
SCodeR achieves new state-of-the-art performance on four code-related tasks over seven datasets.
arXiv Detail & Related papers (2022-10-18T05:17:37Z) - Variational Sparse Coding with Learned Thresholding [6.737133300781134]
We propose a new approach to variational sparse coding that allows us to learn sparse distributions by thresholding samples.
We first evaluate and analyze our method by training a linear generator, showing that it has superior performance, statistical efficiency, and gradient estimation.
arXiv Detail & Related papers (2022-05-07T14:49:50Z) - Addressing Leakage in Self-Supervised Contextualized Code Retrieval [3.693362838682697]
We address contextualized code retrieval, the search for code snippets helpful to fill gaps in a partial input program.
Our approach facilitates a large-scale self-supervised contrastive training by splitting source code randomly into contexts and targets.
To combat leakage between the two, we suggest a novel approach based on mutual identifier masking, dedentation, and the selection of syntax-aligned targets.
arXiv Detail & Related papers (2022-04-17T12:58:38Z) - Extracting Grammars from a Neural Network Parser for Anomaly Detection
in Unknown Formats [79.6676793507792]
Reinforcement learning has recently shown promise as a technique for training an artificial neural network to parse sentences in some unknown format.
This paper presents procedures for extracting production rules from the neural network, and for using these rules to determine whether a given sentence is nominal or anomalous.
arXiv Detail & Related papers (2021-07-30T23:10:24Z) - CIMON: Towards High-quality Hash Codes [63.37321228830102]
We propose a new method named textbfComprehensive stextbfImilarity textbfMining and ctextbfOnsistency leartextbfNing (CIMON)
First, we use global refinement and similarity statistical distribution to obtain reliable and smooth guidance. Second, both semantic and contrastive consistency learning are introduced to derive both disturb-invariant and discriminative hash codes.
arXiv Detail & Related papers (2020-10-15T14:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.