Real-Time GPU-Accelerated Machine Learning Based Multiuser Detection for
5G and Beyond
- URL: http://arxiv.org/abs/2201.05024v2
- Date: Fri, 14 Jan 2022 09:17:35 GMT
- Title: Real-Time GPU-Accelerated Machine Learning Based Multiuser Detection for
5G and Beyond
- Authors: Matthias Mehlhose, Guillermo Marcus, Daniel Sch\"aufele, Daniyal Amir
Awan, Nikolaus Binder, Martin Kasparick, Renato L. G. Cavalcante, S{\l}awomir
Sta\'nczak and Alexander Keller
- Abstract summary: nonlinear beamforming filters can significantly outperform linear approaches in stationary scenarios with massive connectivity.
One of the main challenges comes from the real-time implementation of these algorithms.
This paper explores the acceleration of APSM-based algorithms through massive parallelization.
- Score: 70.81551587109833
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Adaptive partial linear beamforming meets the need of 5G and future 6G
applications for high flexibility and adaptability. Choosing an appropriate
tradeoff between conflicting goals opens the recently proposed multiuser (MU)
detection method. Due to their high spatial resolution, nonlinear beamforming
filters can significantly outperform linear approaches in stationary scenarios
with massive connectivity. However, a dramatic decrease in performance can be
expected in high mobility scenarios because they are very susceptible to
changes in the wireless channel. The robustness of linear filters is required,
considering these changes. One way to respond appropriately is to use online
machine learning algorithms. The theory of algorithms based on the adaptive
projected subgradient method (APSM) is rich, and they promise accurate tracking
capabilities in dynamic wireless environments. However, one of the main
challenges comes from the real-time implementation of these algorithms, which
involve projections on time-varying closed convex sets. While the projection
operations are relatively simple, their vast number poses a challenge in
ultralow latency (ULL) applications where latency constraints must be satisfied
in every radio frame. Taking non-orthogonal multiple access (NOMA) systems as
an example, this paper explores the acceleration of APSM-based algorithms
through massive parallelization. The result is a GPU-accelerated real-time
implementation of an orthogonal frequency-division multiplexing (OFDM)-based
transceiver that enables detection latency of less than one millisecond and
therefore complies with the requirements of 5G and beyond. To meet the
stringent physical layer latency requirements, careful co-design of hardware
and software is essential, especially in virtualized wireless systems with
hardware accelerators.
Related papers
- VARADE: a Variational-based AutoRegressive model for Anomaly Detection on the Edge [7.4646496981460855]
This work presents a novel solution implementing a light autoregressive framework based on variational inference, which is best suited for real-time execution on the edge.
The proposed approach was validated on a robotic arm, part of a pilot production line, and compared with several state-of-the-art algorithms.
arXiv Detail & Related papers (2024-09-23T08:46:15Z) - Online-Score-Aided Federated Learning: Taming the Resource Constraints in Wireless Networks [26.74283774805648]
We propose a new FL algorithm called OSAFL, specifically designed to learn tasks relevant to wireless applications.
Our extensive simulation results on two different tasks -- each with three different datasets -- with four popular ML models validate the effectiveness of OSAFL.
arXiv Detail & Related papers (2024-08-12T01:27:06Z) - Coverage-aware and Reinforcement Learning Using Multi-agent Approach for HD Map QoS in a Realistic Environment [8.853779271331508]
One effective way to optimize the offloading process is by minimizing the transmission time.
This is particularly true in a Vehicular Adhoc Network (VANET) where vehicles frequently download and upload High-definition (HD) map data.
arXiv Detail & Related papers (2024-07-19T12:40:07Z) - Latency-aware Unified Dynamic Networks for Efficient Image Recognition [72.8951331472913]
LAUDNet is a framework to bridge the theoretical and practical efficiency gap in dynamic networks.
It integrates three primary dynamic paradigms-spatially adaptive computation, dynamic layer skipping, and dynamic channel skipping.
It can notably reduce the latency of models like ResNet by over 50% on platforms such as V100,3090, and TX2 GPUs.
arXiv Detail & Related papers (2023-08-30T10:57:41Z) - Latency-aware Spatial-wise Dynamic Networks [33.88843632160247]
We propose a latency-aware spatial-wise dynamic network (LASNet) for deep networks.
LASNet performs coarse-grained spatially adaptive inference under the guidance of a novel latency prediction model.
Experiments on image classification, object detection and instance segmentation demonstrate that the proposed framework significantly improves the practical inference efficiency of deep networks.
arXiv Detail & Related papers (2022-10-12T14:09:27Z) - GPU-Accelerated Machine Learning in Non-Orthogonal Multiple Access [71.58925117604039]
Non-orthogonal multiple access (NOMA) is an interesting technology that enables massive connectivity as required in future 5G and 6G networks.
We propose a neural network architecture that combines the advantages of both linear and non-linear processing.
arXiv Detail & Related papers (2022-06-13T09:38:23Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - Dynamic Network-Assisted D2D-Aided Coded Distributed Learning [59.29409589861241]
We propose a novel device-to-device (D2D)-aided coded federated learning method (D2D-CFL) for load balancing across devices.
We derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time.
Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data.
arXiv Detail & Related papers (2021-11-26T18:44:59Z) - Architecture Aware Latency Constrained Sparse Neural Networks [35.50683537052815]
In this paper, we design an architecture aware latency constrained sparse framework to prune and accelerate CNN models.
We also propose a novel sparse convolution algorithm for efficient computation.
Our system-algorithm co-design framework can achieve much better frontier among network accuracy and latency on resource-constrained mobile devices.
arXiv Detail & Related papers (2021-09-01T03:41:31Z) - Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork [73.94896986868146]
Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
arXiv Detail & Related papers (2021-01-12T08:36:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.