EQ-Net: A Unified Deep Learning Framework for Log-Likelihood Ratio
Estimation and Quantization
- URL: http://arxiv.org/abs/2012.12843v1
- Date: Wed, 23 Dec 2020 18:11:30 GMT
- Title: EQ-Net: A Unified Deep Learning Framework for Log-Likelihood Ratio
Estimation and Quantization
- Authors: Marius Arvinte, Ahmed H. Tewfik, and Sriram Vishwanath
- Abstract summary: We introduce EQ-Net: the first holistic framework that solves both the tasks of log-likelihood ratio (LLR) estimation and quantization using a data-driven method.
We carry out extensive experimental evaluation and demonstrate that our single architecture achieves state-of-the-art results on both tasks.
- Score: 25.484585922608193
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we introduce EQ-Net: the first holistic framework that solves
both the tasks of log-likelihood ratio (LLR) estimation and quantization using
a data-driven method. We motivate our approach with theoretical insights on two
practical estimation algorithms at the ends of the complexity spectrum and
reveal a connection between the complexity of an algorithm and the information
bottleneck method: simpler algorithms admit smaller bottlenecks when
representing their solution. This motivates us to propose a two-stage algorithm
that uses LLR compression as a pretext task for estimation and is focused on
low-latency, high-performance implementations via deep neural networks. We
carry out extensive experimental evaluation and demonstrate that our single
architecture achieves state-of-the-art results on both tasks when compared to
previous methods, with gains in quantization efficiency as high as $20\%$ and
reduced estimation latency by up to $60\%$ when measured on general purpose and
graphical processing units (GPU). In particular, our approach reduces the GPU
inference latency by more than two times in several multiple-input
multiple-output (MIMO) configurations. Finally, we demonstrate that our scheme
is robust to distributional shifts and retains a significant part of its
performance when evaluated on 5G channel models, as well as channel estimation
errors.
Related papers
- YOSO: You-Only-Sample-Once via Compressed Sensing for Graph Neural Network Training [9.02251811867533]
YOSO (You-Only-Sample-Once) is an algorithm designed to achieve efficient training while preserving prediction accuracy.
YOSO not only avoids costly computations in traditional compressed sensing (CS) methods, such as orthonormal basis calculations, but also ensures high-probability accuracy retention.
arXiv Detail & Related papers (2024-11-08T16:47:51Z) - Predicting Probabilities of Error to Combine Quantization and Early Exiting: QuEE [68.6018458996143]
We propose a more general dynamic network that can combine both quantization and early exit dynamic network: QuEE.
Our algorithm can be seen as a form of soft early exiting or input-dependent compression.
The crucial factor of our approach is accurate prediction of the potential accuracy improvement achievable through further computation.
arXiv Detail & Related papers (2024-06-20T15:25:13Z) - Learning the hub graphical Lasso model with the structured sparsity via
an efficient algorithm [1.0923877073891446]
We introduce a two-phase algorithm to estimate hub graphical models.
The proposed algorithm first generates a good initial point via a dual alternating direction method of multipliers.
It then warms a semismooth Newton (SSN) based augmented Lagrangian method (ALM) to compute a solution that is accurate enough for practical tasks.
arXiv Detail & Related papers (2023-08-17T08:24:28Z) - A novel framework for Shot number minimization in Quantum Variational
Algorithms [0.0]
Variational Quantum Algorithms (VQAs) have gained significant attention as a potential solution for various quantum computing applications.
implementing these algorithms on quantum devices often necessitates a substantial number of measurements.
This paper presents a generalized framework for optimization algorithms aiming to reduce the number of shot evaluations in VQAs.
arXiv Detail & Related papers (2023-07-08T19:14:01Z) - Representation Learning with Multi-Step Inverse Kinematics: An Efficient
and Optimal Approach to Rich-Observation RL [106.82295532402335]
Existing reinforcement learning algorithms suffer from computational intractability, strong statistical assumptions, and suboptimal sample complexity.
We provide the first computationally efficient algorithm that attains rate-optimal sample complexity with respect to the desired accuracy level.
Our algorithm, MusIK, combines systematic exploration with representation learning based on multi-step inverse kinematics.
arXiv Detail & Related papers (2023-04-12T14:51:47Z) - ParaFormer: Parallel Attention Transformer for Efficient Feature
Matching [8.552303361149612]
This paper proposes a novel parallel attention model entitled ParaFormer.
It fuses features and keypoint positions through the concept of amplitude and phase, and integrates self- and cross-attention in a parallel manner.
Experiments on various applications, including homography estimation, pose estimation, and image matching, demonstrate that ParaFormer achieves state-of-the-art performance.
The efficient ParaFormer-U variant achieves comparable performance with less than 50% FLOPs of the existing attention-based models.
arXiv Detail & Related papers (2023-03-02T03:29:16Z) - Fully Quantized Image Super-Resolution Networks [81.75002888152159]
We propose a Fully Quantized image Super-Resolution framework (FQSR) to jointly optimize efficiency and accuracy.
We apply our quantization scheme on multiple mainstream super-resolution architectures, including SRResNet, SRGAN and EDSR.
Our FQSR using low bits quantization can achieve on par performance compared with the full-precision counterparts on five benchmark datasets.
arXiv Detail & Related papers (2020-11-29T03:53:49Z) - Temporal Attention-Augmented Graph Convolutional Network for Efficient
Skeleton-Based Human Action Recognition [97.14064057840089]
Graphal networks (GCNs) have been very successful in modeling non-Euclidean data structures.
Most GCN-based action recognition methods use deep feed-forward networks with high computational complexity to process all skeletons in an action.
We propose a temporal attention module (TAM) for increasing the efficiency in skeleton-based action recognition.
arXiv Detail & Related papers (2020-10-23T08:01:55Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Parallelization Techniques for Verifying Neural Networks [52.917845265248744]
We introduce an algorithm based on the verification problem in an iterative manner and explore two partitioning strategies.
We also introduce a highly parallelizable pre-processing algorithm that uses the neuron activation phases to simplify the neural network verification problems.
arXiv Detail & Related papers (2020-04-17T20:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.