Efficient LDPC Decoding using Physical Computation
- URL: http://arxiv.org/abs/2312.02161v1
- Date: Thu, 21 Sep 2023 01:39:07 GMT
- Title: Efficient LDPC Decoding using Physical Computation
- Authors: Uday Kumar Reddy Vengalam, Andrew Hahn, Yongchao Liu, Anshujit Sharma,
Hui Wu, and Michael Huang
- Abstract summary: LDPC decoding could benefit from significant acceleration of physical mechanisms such as Ising machines.
A co-designed Ising machine-based system can improve speed by 3 orders of magnitude.
A physical computation approach can outperform hardwiring state-of-the-art algorithms.
- Score: 7.942478099762508
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Due to 5G deployment, there is significant interest in LDPC decoding. While
much research is devoted on efficient hardwiring of algorithms based on Belief
Propagation (BP), it has been shown that LDPC decoding can be formulated as a
combinatorial optimization problem, which could benefit from significant
acceleration of physical computation mechanisms such as Ising machines. This
approach has so far resulted in poor performance. This paper shows that the
reason is not fundamental but suboptimal hardware and formulation. A
co-designed Ising machine-based system can improve speed by 3 orders of
magnitude. As a result, a physical computation approach can outperform
hardwiring state-of-the-art algorithms. In this paper, we show such an
augmented Ising machine that is 4.4$\times$ more energy efficient than the
state of the art in the literature.
Related papers
- Scalable Thermodynamic Second-order Optimization [0.0]
We propose a scalable algorithm for employing computers to accelerate a popular second-order thermodynamic curvature called Kron-ed approximate curvature (K-FAC)
Numerical experiments show that even under significant quantization noise, the benefits of second-order optimization can be preserved.
We predict substantial speedups for large-scale vision and graph problems based on realistic hardware characteristics.
arXiv Detail & Related papers (2025-02-12T17:44:40Z) - Effcient classical error correction for parity encoded spin systems [0.0]
This paper addresses how to correct errors in a spin readout of PE architecture.
We have shown that independent and identically distributed errors in a spin readout can be corrected by a very simple decoding algorithm.
arXiv Detail & Related papers (2025-02-11T01:31:24Z) - All-to-all reconfigurability with sparse and higher-order Ising machines [0.0]
We introduce a multiplexed architecture that emulates all-to-all network functionality.
We show that running the adaptive parallel tempering algorithm demonstrates competitive algorithmic and prefactor advantages.
scaled magnetic versions of p-bit IMs could lead to orders of magnitude improvements over the state of the art for generic optimization.
arXiv Detail & Related papers (2023-11-21T20:27:02Z) - Hardware Acceleration of Explainable Artificial Intelligence [5.076419064097733]
We propose a simple yet efficient framework to accelerate various XAI algorithms with existing hardware accelerators.
Our proposed approach can lead to real-time outcome interpretation.
arXiv Detail & Related papers (2023-05-04T19:07:29Z) - Harnessing Deep Learning and HPC Kernels via High-Level Loop and Tensor Abstractions on CPU Architectures [67.47328776279204]
This work introduces a framework to develop efficient, portable Deep Learning and High Performance Computing kernels.
We decompose the kernel development in two steps: 1) Expressing the computational core using Processing Primitives (TPPs) and 2) Expressing the logical loops around TPPs in a high-level, declarative fashion.
We demonstrate the efficacy of our approach using standalone kernels and end-to-end workloads that outperform state-of-the-art implementations on diverse CPU platforms.
arXiv Detail & Related papers (2023-04-25T05:04:44Z) - Biologically Plausible Learning on Neuromorphic Hardware Architectures [27.138481022472]
Neuromorphic computing is an emerging paradigm that confronts this imbalance by computations directly in analog memories.
This work is the first to compare the impact of different learning algorithms on Compute-In-Memory-based hardware and vice versa.
arXiv Detail & Related papers (2022-12-29T15:10:59Z) - Batch-efficient EigenDecomposition for Small and Medium Matrices [65.67315418971688]
EigenDecomposition (ED) is at the heart of many computer vision algorithms and applications.
We propose a QR-based ED method dedicated to the application scenarios of computer vision.
arXiv Detail & Related papers (2022-07-09T09:14:12Z) - Efficient Few-Shot Object Detection via Knowledge Inheritance [62.36414544915032]
Few-shot object detection (FSOD) aims at learning a generic detector that can adapt to unseen tasks with scarce training samples.
We present an efficient pretrain-transfer framework (PTF) baseline with no computational increment.
We also propose an adaptive length re-scaling (ALR) strategy to alleviate the vector length inconsistency between the predicted novel weights and the pretrained base weights.
arXiv Detail & Related papers (2022-03-23T06:24:31Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - ALF: Autoencoder-based Low-rank Filter-sharing for Efficient
Convolutional Neural Networks [63.91384986073851]
We propose the autoencoder-based low-rank filter-sharing technique technique (ALF)
ALF shows a reduction of 70% in network parameters, 61% in operations and 41% in execution time, with minimal loss in accuracy.
arXiv Detail & Related papers (2020-07-27T09:01:22Z) - Approximation Algorithms for Sparse Principal Component Analysis [57.5357874512594]
Principal component analysis (PCA) is a widely used dimension reduction technique in machine learning and statistics.
Various approaches to obtain sparse principal direction loadings have been proposed, which are termed Sparse Principal Component Analysis.
We present thresholding as a provably accurate, time, approximation algorithm for the SPCA problem.
arXiv Detail & Related papers (2020-06-23T04:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.