Enabling Real-time Neural Recovery for Cloud Gaming on Mobile Devices
- URL: http://arxiv.org/abs/2307.07847v4
- Date: Mon, 23 Oct 2023 00:51:19 GMT
- Title: Enabling Real-time Neural Recovery for Cloud Gaming on Mobile Devices
- Authors: Zhaoyuan He, Yifan Yang, Shuozhe Li, Diyuan Dai, Lili Qiu, Yuqing Yang
- Abstract summary: We propose a new method for recovering lost or corrupted video frames in cloud gaming.
Unlike traditional video frame recovery, our approach uses game states to significantly enhance recovery accuracy.
We develop a holistic system that consists of (i) efficiently extracting game states, (ii) modifying H.264 video decoder to generate a mask to indicate which portions of video frames need recovery, and (iii) designing a novel neural network to recover either complete or partial video frames.
- Score: 11.530719133935847
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cloud gaming is a multi-billion dollar industry. A client in cloud gaming
sends its movement to the game server on the Internet, which renders and
transmits the resulting video back. In order to provide a good gaming
experience, a latency below 80 ms is required. This means that video rendering,
encoding, transmission, decoding, and display have to finish within that time
frame, which is especially challenging to achieve due to server overload,
network congestion, and losses. In this paper, we propose a new method for
recovering lost or corrupted video frames in cloud gaming. Unlike traditional
video frame recovery, our approach uses game states to significantly enhance
recovery accuracy and utilizes partially decoded frames to recover lost
portions. We develop a holistic system that consists of (i) efficiently
extracting game states, (ii) modifying H.264 video decoder to generate a mask
to indicate which portions of video frames need recovery, and (iii) designing a
novel neural network to recover either complete or partial video frames. Our
approach is extensively evaluated using iPhone 12 and laptop implementations,
and we demonstrate the utility of game states in the game video recovery and
the effectiveness of our overall design.
Related papers
- Real-Time Neural-Enhancement for Online Cloud Gaming [31.971805571638942]
We introduce River, a cloud gaming delivery framework based on the observation that video segment features in cloud gaming are typically repetitive and redundant.
River builds a content-aware encoder that fine-tunes SR models for diverse video segments and stores them in a lookup table.
When delivering cloud gaming video streams online, River checks the video features and retrieves the most relevant SR models to enhance the frame quality.
arXiv Detail & Related papers (2025-01-12T17:28:09Z) - Compressed Domain Prior-Guided Video Super-Resolution for Cloud Gaming Content [39.55748992685093]
We propose a novel lightweight network called Coding Prior-Guided Super-Resolution (CPGSR) to address the SR challenges in compressed game video content.
Inspired by the quantization in video coding, we propose a partitioned focal frequency loss to effectively guide the model's focus on preserving high-frequency information.
arXiv Detail & Related papers (2025-01-03T12:01:36Z) - Large Motion Video Autoencoding with Cross-modal Video VAE [52.13379965800485]
Video Variational Autoencoder (VAE) is essential for reducing video redundancy and facilitating efficient video generation.
Existing Video VAEs have begun to address temporal compression; however, they often suffer from inadequate reconstruction performance.
We present a novel and powerful video autoencoder capable of high-fidelity video encoding.
arXiv Detail & Related papers (2024-12-23T18:58:24Z) - GameIR: A Large-Scale Synthesized Ground-Truth Dataset for Image Restoration over Gaming Content [16.07538127436932]
We develop GameIR, a large-scale computer-synthesized ground-truth dataset to fill in the blanks.
We provide 19200 LR-HR paired ground-truth frames coming from 640 videos rendered at 720p and 1440p for this task.
The second is novel view synthesis (NVS), to support the multiview gaming solution of rendering and transferring part of the multiview frames and generating the remaining frames on the client side.
arXiv Detail & Related papers (2024-08-29T19:11:46Z) - Real-Time Neural Video Recovery and Enhancement on Mobile Devices [15.343787475565836]
We present a novel approach for real-time video enhancement on mobile devices.
We have implemented our approach on an iPhone 12, and it can support 30 frames per second (FPS)
Our approach results in a significant increase in video QoE of 24% - 82% in our video streaming system.
arXiv Detail & Related papers (2023-07-22T19:52:04Z) - Predictive Coding For Animation-Based Video Compression [13.161311799049978]
We propose a predictive coding scheme which uses image animation as a predictor, and codes the residual with respect to the actual target frame.
Our experiments indicate a significant gain, in excess of 70% compared to the HEVC video standard and over 30% compared to VVC.
arXiv Detail & Related papers (2023-07-09T14:40:54Z) - HNeRV: A Hybrid Neural Representation for Videos [56.492309149698606]
Implicit neural representations store videos as neural networks.
We propose a Hybrid Neural Representation for Videos (HNeRV)
With content-adaptive embeddings and re-designed architecture, HNeRV outperforms implicit methods in video regression tasks.
arXiv Detail & Related papers (2023-04-05T17:55:04Z) - ReBotNet: Fast Real-time Video Enhancement [59.08038313427057]
Most restoration networks are slow, have high computational bottleneck, and can't be used for real-time video enhancement.
In this work, we design an efficient and fast framework to perform real-time enhancement for practical use-cases like live video calls and video streams.
To evaluate our method, we emulate two new datasets that real-world video call and streaming scenarios, and show extensive results on multiple datasets where ReBotNet outperforms existing approaches with lower computations, reduced memory requirements, and faster inference time.
arXiv Detail & Related papers (2023-03-23T17:58:05Z) - EVEREST: Efficient Masked Video Autoencoder by Removing Redundant Spatiotemporal Tokens [57.354304637367555]
We present EVEREST, a surprisingly efficient MVA approach for video representation learning.
It finds tokens containing rich motion features and discards uninformative ones during both pre-training and fine-tuning.
Our method significantly reduces the computation and memory requirements of MVA.
arXiv Detail & Related papers (2022-11-19T09:57:01Z) - Transcoded Video Restoration by Temporal Spatial Auxiliary Network [64.63157339057912]
We propose a new method, temporal spatial auxiliary network (TSAN), for transcoded video restoration.
The experimental results demonstrate that the performance of the proposed method is superior to that of the previous techniques.
arXiv Detail & Related papers (2021-12-15T08:10:23Z) - Self-Conditioned Probabilistic Learning of Video Rescaling [70.10092286301997]
We propose a self-conditioned probabilistic framework for video rescaling to learn the paired downscaling and upscaling procedures simultaneously.
We decrease the entropy of the information lost in the downscaling by maximizing its conditioned probability on the strong spatial-temporal prior information.
We extend the framework to a lossy video compression system, in which a gradient estimator for non-differential industrial lossy codecs is proposed.
arXiv Detail & Related papers (2021-07-24T15:57:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.