Improving the List Decoding Version of the Cyclically Equivariant Neural
Decoder
- URL: http://arxiv.org/abs/2106.07964v1
- Date: Tue, 15 Jun 2021 08:37:36 GMT
- Title: Improving the List Decoding Version of the Cyclically Equivariant Neural
Decoder
- Authors: Xiangyu Chen and Min Ye
- Abstract summary: We propose an improved version of the list decoding algorithm for BCH codes and punctured RM codes.
Our new decoder provides up to $2$dB gain over the previous list decoder when measured by BER.
- Score: 33.63188063525036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The cyclically equivariant neural decoder was recently proposed in [Chen-Ye,
International Conference on Machine Learning, 2021] to decode cyclic codes. In
the same paper, a list decoding procedure was also introduced for two widely
used classes of cyclic codes -- BCH codes and punctured Reed-Muller (RM) codes.
While the list decoding procedure significantly improves the Frame Error Rate
(FER) of the cyclically equivariant neural decoder, the Bit Error Rate (BER) of
the list decoding procedure is even worse than the unique decoding algorithm
when the list size is small. In this paper, we propose an improved version of
the list decoding algorithm for BCH codes and punctured RM codes. Our new
proposal significantly reduces the BER while maintaining the same (in some
cases even smaller) FER. More specifically, our new decoder provides up to
$2$dB gain over the previous list decoder when measured by BER, and the running
time of our new decoder is $15\%$ smaller. Code available at
https://github.com/improvedlistdecoder/code
Related papers
- Decoding Quasi-Cyclic Quantum LDPC Codes [23.22566380210149]
Quantum low-density parity-check (qLDPC) codes are an important component in the quest for fault tolerance.
Recent progress on qLDPC codes has led to constructions which are quantumally good, and which admit linear-time decoders to correct errors affecting a constant fraction of codeword qubits.
In practice, the surface/toric codes, which are the product of two repetition codes, are still often the qLDPC codes of choice.
arXiv Detail & Related papers (2024-11-07T06:25:27Z) - Collective Bit Flipping-Based Decoding of Quantum LDPC Codes [0.6554326244334866]
We improve both the error correction performance and decoding latency of variable degree-3 (dv-3) QLDPC codes under iterative decoding.
Our decoding scheme is based on applying a modified version of bit flipping (BF) decoding, namely two-bit bit flipping (TBF) decoding.
arXiv Detail & Related papers (2024-06-24T18:51:48Z) - Learning Linear Block Error Correction Codes [62.25533750469467]
We propose for the first time a unified encoder-decoder training of binary linear block codes.
We also propose a novel Transformer model in which the self-attention masking is performed in a differentiable fashion for the efficient backpropagation of the code gradient.
arXiv Detail & Related papers (2024-05-07T06:47:12Z) - Progressive-Proximity Bit-Flipping for Decoding Surface Codes [8.971989179518214]
Topological quantum codes, such as toric and surface codes, are excellent candidates for hardware implementation.
Existing decoders often fall short of meeting requirements such as having low computational complexity.
We propose a novel bit-flipping (BF) decoder tailored for toric and surface codes.
arXiv Detail & Related papers (2024-02-24T22:38:05Z) - Bit-flipping Decoder Failure Rate Estimation for (v,w)-regular Codes [84.0257274213152]
We propose a new technique to provide accurate estimates of the DFR of a two-iterations (parallel) bit flipping decoder.
We validate our results, providing comparisons of the modeled and simulated weight of the syndrome, incorrectly-guessed error bit distribution at the end of the first iteration, and two-itcrypteration Decoding Failure Rates (DFR)
arXiv Detail & Related papers (2024-01-30T11:40:24Z) - Machine Learning-Aided Efficient Decoding of Reed-Muller Subcodes [59.55193427277134]
Reed-Muller (RM) codes achieve the capacity of general binary-input memoryless symmetric channels.
RM codes only admit limited sets of rates.
Efficient decoders are available for RM codes at finite lengths.
arXiv Detail & Related papers (2023-01-16T04:11:14Z) - Graph Neural Networks for Channel Decoding [71.15576353630667]
We showcase competitive decoding performance for various coding schemes, such as low-density parity-check (LDPC) and BCH codes.
The idea is to let a neural network (NN) learn a generalized message passing algorithm over a given graph.
We benchmark our proposed decoder against state-of-the-art in conventional channel decoding as well as against recent deep learning-based results.
arXiv Detail & Related papers (2022-07-29T15:29:18Z) - Cyclically Equivariant Neural Decoders for Cyclic Codes [33.63188063525036]
We propose a novel neural decoder for cyclic codes by exploiting their cyclically invariant property.
Our new decoder consistently outperforms previous neural decoders when decoding cyclic codes.
Finally, we propose a list decoding procedure that can significantly reduce the decoding error probability for BCH codes and punctured RM codes.
arXiv Detail & Related papers (2021-05-12T09:41:13Z) - Pruning Neural Belief Propagation Decoders [77.237958592189]
We introduce a method to tailor an overcomplete parity-check matrix to (neural) BP decoding using machine learning.
We achieve performance within 0.27 dB and 1.5 dB of the ML performance while reducing the complexity of the decoder.
arXiv Detail & Related papers (2020-01-21T12:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.