Network state Estimation using Raw Video Analysis: vQoS-GAN based
non-intrusive Deep Learning Approach
- URL: http://arxiv.org/abs/2204.07062v1
- Date: Tue, 22 Mar 2022 10:42:19 GMT
- Title: Network state Estimation using Raw Video Analysis: vQoS-GAN based
non-intrusive Deep Learning Approach
- Authors: Renith G, Harikrishna Warrier, Yogesh Gupta
- Abstract summary: vQoS GAN can estimate the network state parameters from the degraded received video data.
A robust and unique design of deep learning network model has been trained with the video data along with data rate and packet loss class labels.
The proposed semi supervised generative adversarial network can additionally reconstruct the degraded video data to its original form for a better end user experience.
- Score: 5.8010446129208155
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Content based providers transmits real time complex signal such as video data
from one region to another. During this transmission process, the signals
usually end up distorted or degraded where the actual information present in
the video is lost. This normally happens in the streaming video services
applications. Hence there is a need to know the level of degradation that
happened in the receiver side. This video degradation can be estimated by
network state parameters like data rate and packet loss values. Our proposed
solution vQoS GAN (video Quality of Service Generative Adversarial Network) can
estimate the network state parameters from the degraded received video data
using a deep learning approach of semi supervised generative adversarial
network algorithm. A robust and unique design of deep learning network model
has been trained with the video data along with data rate and packet loss class
labels and achieves over 95 percent of training accuracy. The proposed semi
supervised generative adversarial network can additionally reconstruct the
degraded video data to its original form for a better end user experience.
Related papers
- CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - STIP: A SpatioTemporal Information-Preserving and Perception-Augmented
Model for High-Resolution Video Prediction [78.129039340528]
We propose a Stemporal Information-Preserving and Perception-Augmented Model (STIP) to solve the above two problems.
The proposed model aims to preserve thetemporal information for videos during the feature extraction and the state transitions.
Experimental results show that the proposed STIP can predict videos with more satisfactory visual quality compared with a variety of state-of-the-art methods.
arXiv Detail & Related papers (2022-06-09T09:49:04Z) - VPN: Video Provenance Network for Robust Content Attribution [72.12494245048504]
We present VPN - a content attribution method for recovering provenance information from videos shared online.
We learn a robust search embedding for matching such video, using full-length or truncated video queries.
Once matched against a trusted database of video clips, associated information on the provenance of the clip is presented to the user.
arXiv Detail & Related papers (2021-09-21T09:07:05Z) - SignalNet: A Low Resolution Sinusoid Decomposition and Estimation
Network [79.04274563889548]
We propose SignalNet, a neural network architecture that detects the number of sinusoids and estimates their parameters from quantized in-phase and quadrature samples.
We introduce a worst-case learning threshold for comparing the results of our network relative to the underlying data distributions.
In simulation, we find that our algorithm is always able to surpass the threshold for three-bit data but often cannot exceed the threshold for one-bit data.
arXiv Detail & Related papers (2021-06-10T04:21:20Z) - Self-Supervised Adaptation for Video Super-Resolution [7.26562478548988]
Single-image super-resolution (SISR) networks can adapt their network parameters to specific input images.
We present a new learning algorithm that allows conventional video super-resolution (VSR) networks to adapt their parameters to test video frames.
arXiv Detail & Related papers (2021-03-18T08:30:24Z) - Ultra-low bitrate video conferencing using deep image animation [7.263312285502382]
We propose a novel deep learning approach for ultra-low video compression for video conferencing applications.
We employ deep neural networks to encode motion information as keypoint displacement and reconstruct the video signal at the decoder side.
arXiv Detail & Related papers (2020-12-01T09:06:34Z) - Training Strategies and Data Augmentations in CNN-based DeepFake Video
Detection [17.696134665850447]
The accuracy of automated systems for face forgery detection in videos is still quite limited and generally biased toward the dataset used to design and train a specific detection system.
In this paper we analyze how different training strategies and data augmentation techniques affect CNN-based deepfake detectors when training and testing on the same dataset or across different datasets.
arXiv Detail & Related papers (2020-11-16T08:50:56Z) - HRVGAN: High Resolution Video Generation using Spatio-Temporal GAN [0.0]
We present a novel network for high resolution video generation.
Our network uses ideas from Wasserstein GANs by enforcing k-Lipschitz constraint on the loss term and Conditional GANs using class labels for training and testing.
arXiv Detail & Related papers (2020-08-17T20:45:59Z) - Auto-Rectify Network for Unsupervised Indoor Depth Estimation [119.82412041164372]
We establish that the complex ego-motions exhibited in handheld settings are a critical obstacle for learning depth.
We propose a data pre-processing method that rectifies training images by removing their relative rotations for effective learning.
Our results outperform the previous unsupervised SOTA method by a large margin on the challenging NYUv2 dataset.
arXiv Detail & Related papers (2020-06-04T08:59:17Z) - Learning the Loss Functions in a Discriminative Space for Video
Restoration [48.104095018697556]
We propose a new framework for building effective loss functions by learning a discriminative space specific to a video restoration task.
Our framework is similar to GANs in that we iteratively train two networks - a generator and a loss network.
Experiments on video superresolution and deblurring show that our method generates visually more pleasing videos.
arXiv Detail & Related papers (2020-03-20T06:58:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.