RELDEC: Reinforcement Learning-Based Decoding of Moderate Length LDPC
Codes
- URL: http://arxiv.org/abs/2112.13934v3
- Date: Thu, 27 Jul 2023 16:53:49 GMT
- Title: RELDEC: Reinforcement Learning-Based Decoding of Moderate Length LDPC
Codes
- Authors: Salman Habib, Allison Beemer, and Joerg Kliewer
- Abstract summary: RELDEC is a novel approach for sequential decoding of moderate length low-density parity-check (LDPC) codes.
An optimized decoding policy is obtained via reinforcement learning based on a Markov decision process (MDP)
The proposed RELDEC scheme significantly outperforms standard flooding and random sequential decoding for a variety of LDPC codes.
- Score: 4.588028371034406
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we propose RELDEC, a novel approach for sequential decoding of
moderate length low-density parity-check (LDPC) codes. The main idea behind
RELDEC is that an optimized decoding policy is subsequently obtained via
reinforcement learning based on a Markov decision process (MDP). In contrast to
our previous work, where an agent learns to schedule only a single check node
(CN) within a group (cluster) of CNs per iteration, in this work we train the
agent to schedule all CNs in a cluster, and all clusters in every iteration.
That is, in each learning step of RELDEC an agent learns to schedule CN
clusters sequentially depending on a reward associated with the outcome of
scheduling a particular cluster. We also modify the state space representation
of the MDP, enabling RELDEC to be suitable for larger block length LDPC codes
than those studied in our previous work. Furthermore, to address decoding under
varying channel conditions, we propose agile meta-RELDEC (AM-RELDEC) that
employs meta-reinforcement learning. The proposed RELDEC scheme significantly
outperforms standard flooding and random sequential decoding for a variety of
LDPC codes, including codes designed for 5G new radio.
Related papers
- Action-List Reinforcement Learning Syndrome Decoding for Binary Linear Block Codes [3.3148826359547514]
We describe the methodology for mapping the iterative decoding process into Markov Decision Processes (MDPs)<n>A truncated MDP is proposed to reduce the number of states in the MDP by learning a Hamming ball with a specified radius around codewords.<n>We design an action-list decoder based on the Deep-Q network values that substantially enhance performance.
arXiv Detail & Related papers (2025-07-23T19:42:51Z) - On-Policy Optimization with Group Equivalent Preference for Multi-Programming Language Understanding [5.429445008970627]
Large language models (LLMs) achieve remarkable performance in code generation tasks.<n>A significant performance disparity persists between popular programming languages.<n>We leverage the code translation task to train LLMs, thereby facilitating the transfer of coding proficiency.
arXiv Detail & Related papers (2025-05-19T05:25:29Z) - Collab: Controlled Decoding using Mixture of Agents for LLM Alignment [90.6117569025754]
Reinforcement learning from human feedback has emerged as an effective technique to align Large Language models.
Controlled Decoding provides a mechanism for aligning a model at inference time without retraining.
We propose a mixture of agent-based decoding strategies leveraging the existing off-the-shelf aligned LLM policies.
arXiv Detail & Related papers (2025-03-27T17:34:25Z) - Goal-Oriented Source Coding using LDPC Codes for Compressed-Domain Image Classification [3.2657732635702375]
This paper investigates the use of low-density parity-check (LDPC) codes as an alternative entropy-coding approach.
It is hypothesized that the structured nature of LDPC codes can be leveraged more effectively by deep learning models for tasks like classification.
Experiments on datasets like MNIST, Fashion-MNIST, and CIFAR show that LDPC codes outperform Huffman and Arithmetic coding in classification tasks.
arXiv Detail & Related papers (2025-03-15T01:52:09Z) - Threshold Selection for Iterative Decoding of $(v,w)$-regular Binary Codes [84.0257274213152]
Iterative bit flipping decoders are an efficient choice for sparse $(v,w)$-regular codes.
We propose concrete criteria for threshold determination, backed by a closed form model.
arXiv Detail & Related papers (2025-01-23T17:38:22Z) - GNN-based Auto-Encoder for Short Linear Block Codes: A DRL Approach [43.17241175857862]
It integrates deep reinforcement learning (DRL) and graph neural networks (GNN) in code design.
An edge-weighted GNN (EW-GNN) decoder is proposed, which operates on the Tanner graph with an iterative message-passing structure.
An iterative joint training of the DRL-based code designer and the EW-GNN decoder is performed to optimize the end-end encoding and decoding process.
arXiv Detail & Related papers (2024-12-03T00:25:14Z) - Decoding Quantum LDPC Codes Using Graph Neural Networks [52.19575718707659]
We propose a novel decoding method for Quantum Low-Density Parity-Check (QLDPC) codes based on Graph Neural Networks (GNNs)
The proposed GNN-based QLDPC decoder exploits the sparse graph structure of QLDPC codes and can be implemented as a message-passing decoding algorithm.
arXiv Detail & Related papers (2024-08-09T16:47:49Z) - Learning on JPEG-LDPC Compressed Images: Classifying with Syndromes [3.2657732635702375]
In goal-oriented communications, the objective of the receiver is often to apply a Deep-Learning model, rather than reconstructing the original data.
We propose an alternative approach in which entropic coding is realized with Low-Density Parity Check (LDPC) codes.
arXiv Detail & Related papers (2024-03-15T11:07:38Z) - StepCoder: Improve Code Generation with Reinforcement Learning from
Compiler Feedback [58.20547418182074]
We introduce StepCoder, a novel framework for code generation, consisting of two main components.
CCCS addresses the exploration challenge by breaking the long sequences code generation task into a Curriculum of Code Completion Subtasks.
FGO only optimize the model by masking the unexecuted code segments to provide Fine-Grained Optimization.
Our method improves the ability to explore the output space and outperforms state-of-the-art approaches in corresponding benchmarks.
arXiv Detail & Related papers (2024-02-02T13:14:31Z) - Graph Neural Networks for Enhanced Decoding of Quantum LDPC Codes [6.175503577352742]
We propose a differentiable iterative decoder for quantum low-density parity-check (LDPC) codes.
The proposed algorithm is composed of classical belief propagation (BP) decoding stages and intermediate graph neural network (GNN) layers.
arXiv Detail & Related papers (2023-10-26T19:56:25Z) - Exploring Continual Learning for Code Generation Models [80.78036093054855]
Continual Learning (CL) is an important aspect that remains underexplored in the code domain.
We introduce a benchmark called CodeTask-CL that covers a wide range of tasks, including code generation, translation, summarization, and refinement.
We find that effective methods like Prompt Pooling (PP) suffer from catastrophic forgetting due to the unstable training of the prompt selection mechanism.
arXiv Detail & Related papers (2023-07-05T16:58:39Z) - Deep Joint Source-Channel Coding with Iterative Source Error Correction [11.41076729592696]
We propose an iterative source error correction (ISEC) decoding scheme for deep-learning-based joint source-channel code (Deep J SCC)
Given a noisyword received through the channel, we use a Deep J SCC encoder and decoder pair to update the code iteratively.
The proposed scheme produces more reliable source reconstruction results compared to the baseline when the channel noise characteristics do not match the ones used during training.
arXiv Detail & Related papers (2023-02-17T22:50:58Z) - Machine Learning-Aided Efficient Decoding of Reed-Muller Subcodes [59.55193427277134]
Reed-Muller (RM) codes achieve the capacity of general binary-input memoryless symmetric channels.
RM codes only admit limited sets of rates.
Efficient decoders are available for RM codes at finite lengths.
arXiv Detail & Related papers (2023-01-16T04:11:14Z) - Graph Neural Networks for Channel Decoding [71.15576353630667]
We showcase competitive decoding performance for various coding schemes, such as low-density parity-check (LDPC) and BCH codes.
The idea is to let a neural network (NN) learn a generalized message passing algorithm over a given graph.
We benchmark our proposed decoder against state-of-the-art in conventional channel decoding as well as against recent deep learning-based results.
arXiv Detail & Related papers (2022-07-29T15:29:18Z) - Coding for Distributed Multi-Agent Reinforcement Learning [12.366967700730449]
Stragglers arise frequently in a distributed learning system, due to the existence of various system disturbances.
We propose a coded distributed learning framework, which speeds up the training of MARL algorithms in the presence of stragglers.
Different coding schemes, including maximum distance separable (MDS)code, random sparse code, replication-based code, and regular low density parity check (LDPC) code are also investigated.
arXiv Detail & Related papers (2021-01-07T00:22:34Z) - Pruning Neural Belief Propagation Decoders [77.237958592189]
We introduce a method to tailor an overcomplete parity-check matrix to (neural) BP decoding using machine learning.
We achieve performance within 0.27 dB and 1.5 dB of the ML performance while reducing the complexity of the decoder.
arXiv Detail & Related papers (2020-01-21T12:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.