Asynchronous Hybrid Reinforcement Learning for Latency and Reliability
Optimization in the Metaverse over Wireless Communications
- URL: http://arxiv.org/abs/2212.14749v1
- Date: Fri, 30 Dec 2022 14:40:00 GMT
- Title: Asynchronous Hybrid Reinforcement Learning for Latency and Reliability
Optimization in the Metaverse over Wireless Communications
- Authors: Wenhan Yu, Terence Jie Chua, Jun Zhao
- Abstract summary: Real-time digital twinning of real-world scenes is increasing.
The disparity in transmitted scene dimension (2D as opposed to 3D) leads to asymmetric data sizes in uplink (UL) and downlink (DL)
We design a novel multi-agent reinforcement learning algorithm structure, namely Asynchronous Actors Hybrid Critic (AAHC)
- Score: 8.513938423514636
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Technology advancements in wireless communications and high-performance
Extended Reality (XR) have empowered the developments of the Metaverse. The
demand for Metaverse applications and hence, real-time digital twinning of
real-world scenes is increasing. Nevertheless, the replication of 2D physical
world images into 3D virtual world scenes is computationally intensive and
requires computation offloading. The disparity in transmitted scene dimension
(2D as opposed to 3D) leads to asymmetric data sizes in uplink (UL) and
downlink (DL). To ensure the reliability and low latency of the system, we
consider an asynchronous joint UL-DL scenario where in the UL stage, the
smaller data size of the physical world scenes captured by multiple extended
reality users (XUs) will be uploaded to the Metaverse Console (MC) to be
construed and rendered. In the DL stage, the larger-size 3D virtual world
scenes need to be transmitted back to the XUs. The decisions pertaining to
computation offloading and channel assignment are optimized in the UL stage,
and the MC will optimize power allocation for users assigned with a channel in
the UL transmission stage. Some problems arise therefrom: (i) interactive
multi-process chain, specifically Asynchronous Markov Decision Process (AMDP),
(ii) joint optimization in multiple processes, and (iii) high-dimensional
objective functions, or hybrid reward scenarios. To ensure the reliability and
low latency of the system, we design a novel multi-agent reinforcement learning
algorithm structure, namely Asynchronous Actors Hybrid Critic (AAHC). Extensive
experiments demonstrate that compared to proposed baselines, AAHC obtains
better solutions with preferable training time.
Related papers
- ALOcc: Adaptive Lifting-based 3D Semantic Occupancy and Cost Volume-based Flow Prediction [89.89610257714006]
Existing methods prioritize higher accuracy to cater to the demands of these tasks.
We introduce a series of targeted improvements for 3D semantic occupancy prediction and flow estimation.
Our purelytemporalal architecture framework, named ALOcc, achieves an optimal tradeoff between speed and accuracy.
arXiv Detail & Related papers (2024-11-12T11:32:56Z) - Towards 3D Semantic Scene Completion for Autonomous Driving: A Meta-Learning Framework Empowered by Deformable Large-Kernel Attention and Mamba Model [1.6835437621159244]
We introduce MetaSSC, a novel meta-learning-based framework for semantic scene completion (SSC)
Our approach begins with a voxel-based semantic segmentation (SS) pretraining task, aimed at exploring the semantics and geometry of incomplete regions.
Using simulated cooperative perception datasets, we supervise the perception training of a single vehicle using aggregated sensor data.
This meta-knowledge is then adapted to the target domain through a dual-phase training strategy, enabling efficient deployment.
arXiv Detail & Related papers (2024-11-06T05:11:25Z) - Hyperdimensional Computing Empowered Federated Foundation Model over Wireless Networks for Metaverse [56.384390765357004]
We propose an integrated federated split learning and hyperdimensional computing framework for emerging foundation models.
This novel approach reduces communication costs, computation load, and privacy risks, making it suitable for resource-constrained edge devices in the Metaverse.
arXiv Detail & Related papers (2024-08-26T17:03:14Z) - RAVEN: Rethinking Adversarial Video Generation with Efficient Tri-plane Networks [93.18404922542702]
We present a novel video generative model designed to address long-term spatial and temporal dependencies.
Our approach incorporates a hybrid explicit-implicit tri-plane representation inspired by 3D-aware generative frameworks.
Our model synthesizes high-fidelity video clips at a resolution of $256times256$ pixels, with durations extending to more than $5$ seconds at a frame rate of 30 fps.
arXiv Detail & Related papers (2024-01-11T16:48:44Z) - Federated Multi-View Synthesizing for Metaverse [52.59476179535153]
The metaverse is expected to provide immersive entertainment, education, and business applications.
Virtual reality (VR) transmission over wireless networks is data- and computation-intensive.
We have developed a novel multi-view synthesizing framework that can efficiently provide synthesizing, storage, and communication resources for wireless content delivery in the metaverse.
arXiv Detail & Related papers (2023-12-18T13:51:56Z) - Mutual Information-driven Triple Interaction Network for Efficient Image
Dehazing [54.168567276280505]
We propose a novel Mutual Information-driven Triple interaction Network (MITNet) for image dehazing.
The first stage, named amplitude-guided haze removal, aims to recover the amplitude spectrum of the hazy images for haze removal.
The second stage, named phase-guided structure refined, devotes to learning the transformation and refinement of the phase spectrum.
arXiv Detail & Related papers (2023-08-14T08:23:58Z) - Dynamic Network-Assisted D2D-Aided Coded Distributed Learning [59.29409589861241]
We propose a novel device-to-device (D2D)-aided coded federated learning method (D2D-CFL) for load balancing across devices.
We derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time.
Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data.
arXiv Detail & Related papers (2021-11-26T18:44:59Z) - Real-time 3D Facial Tracking via Cascaded Compositional Learning [30.660564667452118]
We learn a cascade of globally-optimized modular boosted ferns (GoMBF) to solve multi-modal facial motion regression for real-time 3D facial tracking from a monocular RGB camera.
GoMBF is a deep composition of multiple regression models with each is a boosted ferns initially trained to predict partial motion parameters of the same modality.
arXiv Detail & Related papers (2020-09-02T10:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.