NeuSaver: Neural Adaptive Power Consumption Optimization for Mobile
Video Streaming
- URL: http://arxiv.org/abs/2107.07127v1
- Date: Thu, 15 Jul 2021 05:17:17 GMT
- Title: NeuSaver: Neural Adaptive Power Consumption Optimization for Mobile
Video Streaming
- Authors: Kyoungjun Park, Myungchul Kim, Laihyuk Park
- Abstract summary: NeuSaver applies an adaptive frame rate to each video chunk without compromising user experience.
NeuSaver generates an optimal policy that determines the appropriate frame rate for each video chunk.
NeuSaver effectively reduces the power consumption of mobile devices when streaming video by an average of 16.14% and up to 23.12% while achieving high QoE.
- Score: 3.3194866396158003
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Video streaming services strive to support high-quality videos at higher
resolutions and frame rates to improve the quality of experience (QoE).
However, high-quality videos consume considerable amounts of energy on mobile
devices. This paper proposes NeuSaver, which reduces the power consumption of
mobile devices when streaming videos by applying an adaptive frame rate to each
video chunk without compromising user experience. NeuSaver generates an optimal
policy that determines the appropriate frame rate for each video chunk using
reinforcement learning (RL). The RL model automatically learns the policy that
maximizes the QoE goals based on previous observations. NeuSaver also uses an
asynchronous advantage actor-critic algorithm to reinforce the RL model quickly
and robustly. Streaming servers that support NeuSaver preprocesses videos into
segments with various frame rates, which is similar to the process of creating
videos with multiple bit rates in dynamic adaptive streaming over HTTP.
NeuSaver utilizes the commonly used H.264 video codec. We evaluated NeuSaver in
various experiments and a user study through four video categories along with
the state-of-the-art model. Our experiments showed that NeuSaver effectively
reduces the power consumption of mobile devices when streaming video by an
average of 16.14% and up to 23.12% while achieving high QoE.
Related papers
- Plug-and-Play Versatile Compressed Video Enhancement [57.62582951699999]
Video compression effectively reduces the size of files, making it possible for real-time cloud computing.
However, it comes at the cost of visual quality, challenges the robustness of downstream vision models.
We present a versatile-aware enhancement framework that adaptively enhance videos under different compression settings.
arXiv Detail & Related papers (2025-04-21T18:39:31Z) - Streaming Video Question-Answering with In-context Video KV-Cache Retrieval [10.990431921021585]
We propose ReKV, a training-free approach that enables efficient streaming video question-answering (StreamingVQA)
Our approach analyzes long videos in a streaming manner, allowing for prompt responses as soon as user queries are received.
arXiv Detail & Related papers (2025-03-01T15:53:33Z) - Adaptive Caching for Faster Video Generation with Diffusion Transformers [52.73348147077075]
Diffusion Transformers (DiTs) rely on larger models and heavier attention mechanisms, resulting in slower inference speeds.
We introduce a training-free method to accelerate video DiTs, termed Adaptive Caching (AdaCache)
We also introduce a Motion Regularization (MoReg) scheme to utilize video information within AdaCache, controlling the compute allocation based on motion content.
arXiv Detail & Related papers (2024-11-04T18:59:44Z) - LL-GABR: Energy Efficient Live Video Streaming Using Reinforcement
Learning [3.360922672565235]
We propose LLGABR, a deep reinforcement learning approach that models the QoE using perceived video quality instead of energy consumption.
We show that LLGABR outperforms the state-of-the-art approaches by up to 44% in terms of perceptual QoE and a 73% increase in energy efficiency.
arXiv Detail & Related papers (2024-02-14T18:43:19Z) - NU-Class Net: A Novel Approach for Video Quality Enhancement [1.7763979745248648]
This paper introduces NU-Class Net, an innovative deep-learning model designed to mitigate compression artifacts stemming from lossy compression codecs.
By employing the NU-Class Net, the video encoder within the video-capturing node can reduce output quality, thereby generating low-bit-rate videos.
Experimental results affirm the efficacy of the proposed model in enhancing the perceptible quality of videos, especially those streamed at low bit rates.
arXiv Detail & Related papers (2024-01-02T11:46:42Z) - Towards High-Quality and Efficient Video Super-Resolution via
Spatial-Temporal Data Overfitting [27.302681897961588]
Deep convolutional neural networks (DNNs) are widely used in various fields of computer vision.
We propose a novel method for high-quality and efficient video resolution upscaling tasks.
We deploy our models on an off-the-shelf mobile phone, and experimental results show that our method achieves real-time video super-resolution with high video quality.
arXiv Detail & Related papers (2023-03-15T02:40:02Z) - Rate-Perception Optimized Preprocessing for Video Coding [15.808458228130261]
We propose a rate-perception optimized preprocessing (RPP) method to improve the rate-distortion performance.
Our RPP method is very simple and efficient which is not required any changes in the setting of video encoding, streaming, and decoding.
In our subjective visual quality test, 87% of users think videos with RPP are better or equal to videos by only using the to compress these videos with RPP save about 12%.
arXiv Detail & Related papers (2023-01-25T08:21:52Z) - Deep Unsupervised Key Frame Extraction for Efficient Video
Classification [63.25852915237032]
This work presents an unsupervised method to retrieve the key frames, which combines Convolutional Neural Network (CNN) and Temporal Segment Density Peaks Clustering (TSDPC)
The proposed TSDPC is a generic and powerful framework and it has two advantages compared with previous works, one is that it can calculate the number of key frames automatically.
Furthermore, a Long Short-Term Memory network (LSTM) is added on the top of the CNN to further elevate the performance of classification.
arXiv Detail & Related papers (2022-11-12T20:45:35Z) - OCSampler: Compressing Videos to One Clip with Single-step Sampling [82.0417131211353]
We propose a framework named OCSampler to explore a compact yet effective video representation with one short clip.
Our basic motivation is that the efficient video recognition task lies in processing a whole sequence at once rather than picking up frames sequentially.
arXiv Detail & Related papers (2022-01-12T09:50:38Z) - FAVER: Blind Quality Prediction of Variable Frame Rate Videos [47.951054608064126]
Video quality assessment (VQA) remains an important and challenging problem that affects many applications at the widest scales.
We propose a first-of-a-kind blind VQA model for evaluating HFR videos, which we dub the Framerate-Aware Video Evaluator w/o Reference (FAVER)
Our experiments on several HFR video quality datasets show that FAVER outperforms other blind VQA algorithms at a reasonable computational cost.
arXiv Detail & Related papers (2022-01-05T07:54:12Z) - Subjective and Objective Quality Assessment of High Frame Rate Videos [60.970191379802095]
High frame rate (HFR) videos are becoming increasingly common with the tremendous popularity of live, high-action streaming content such as sports.
Live-YT-HFR dataset is comprised of 480 videos having 6 different frame rates, obtained from 16 diverse contents.
To obtain subjective labels on the videos, we conducted a human study yielding 19,000 human quality ratings obtained from a pool of 85 human subjects.
arXiv Detail & Related papers (2020-07-22T19:11:42Z) - Capturing Video Frame Rate Variations via Entropic Differencing [63.749184706461826]
We propose a novel statistical entropic differencing method based on a Generalized Gaussian Distribution model.
Our proposed model correlates very well with subjective scores in the recently proposed LIVE-YT-HFR database.
arXiv Detail & Related papers (2020-06-19T22:16:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.