GRACE: Loss-Resilient Real-Time Video through Neural Codecs
- URL: http://arxiv.org/abs/2305.12333v4
- Date: Tue, 12 Mar 2024 21:40:53 GMT
- Title: GRACE: Loss-Resilient Real-Time Video through Neural Codecs
- Authors: Yihua Cheng, Ziyi Zhang, Hanchen Li, Anton Arapin, Yue Zhang, Qizheng
Zhang, Yuhan Liu, Xu Zhang, Francis Y. Yan, Amrita Mazumdar, Nick Feamster,
Junchen Jiang
- Abstract summary: In real-time video communication, retransmitting lost packets over high-latency networks is not viable due to strict latency requirements.
We present a loss-resilient real-time video system called GRACE, which preserves the user's quality of experience (QE) across a wide range of packet losses.
- Score: 31.006987868475683
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In real-time video communication, retransmitting lost packets over
high-latency networks is not viable due to strict latency requirements. To
counter packet losses without retransmission, two primary strategies are
employed -- encoder-based forward error correction (FEC) and decoder-based
error concealment. The former encodes data with redundancy before transmission,
yet determining the optimal redundancy level in advance proves challenging. The
latter reconstructs video from partially received frames, but dividing a frame
into independently coded partitions inherently compromises compression
efficiency, and the lost information cannot be effectively recovered by the
decoder without adapting the encoder. We present a loss-resilient real-time
video system called GRACE, which preserves the user's quality of experience
(QoE) across a wide range of packet losses through a new neural video codec.
Central to GRACE's enhanced loss resilience is its joint training of the neural
encoder and decoder under a spectrum of simulated packet losses. In lossless
scenarios, GRACE achieves video quality on par with conventional codecs (e.g.,
H.265). As the loss rate escalates, GRACE exhibits a more graceful, less
pronounced decline in quality, consistently outperforming other loss-resilient
schemes. Through extensive evaluation on various videos and real network
traces, we demonstrate that GRACE reduces undecodable frames by 95% and stall
duration by 90% compared with FEC, while markedly boosting video quality over
error concealment methods. In a user study with 240 crowdsourced participants
and 960 subjective ratings, GRACE registers a 38% higher mean opinion score
(MOS) than other baselines.
Related papers
- Lightweight Hybrid Video Compression Framework Using Reference-Guided
Restoration Network [12.033330902093152]
We propose a new lightweight hybrid video consisting of a conventional video(HEVC / VVC), a lossless image, and our new restoration network.
Our method achieves comparable performance to top-tier methods, even when applied to HEVC.
arXiv Detail & Related papers (2023-03-21T04:42:44Z) - Pruned Lightweight Encoders for Computer Vision [0.0]
We show that ASTC and JPEG XS encoding configurations can be used on a near-sensor edge device to ensure low latency.
We reduced the classification accuracy and segmentation mean over union (mIoU) degradation due to ASTC compression to 4.9-5.0 percentage points (pp) and 4.4-4.0 pp, respectively.
In terms of encoding speed, our ASTC encoder implementation is 2.3x faster than JPEG.
arXiv Detail & Related papers (2022-11-23T17:11:48Z) - High Fidelity Neural Audio Compression [92.4812002532009]
We introduce a state-of-the-art real-time, high-fidelity, audio leveraging neural networks.
It consists in a streaming encoder-decoder architecture with quantized latent space trained in an end-to-end fashion.
We simplify and speed-up the training by using a single multiscale spectrogram adversary.
arXiv Detail & Related papers (2022-10-24T17:52:02Z) - Deep Lossy Plus Residual Coding for Lossless and Near-lossless Image
Compression [85.93207826513192]
We propose a unified and powerful deep lossy plus residual (DLPR) coding framework for both lossless and near-lossless image compression.
We solve the joint lossy and residual compression problem in the approach of VAEs.
In the near-lossless mode, we quantize the original residuals to satisfy a given $ell_infty$ error bound.
arXiv Detail & Related papers (2022-09-11T12:11:56Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - A Coding Framework and Benchmark towards Low-Bitrate Video Understanding [63.05385140193666]
We propose a traditional-neural mixed coding framework that takes advantage of both traditional codecs and neural networks (NNs)
The framework is optimized by ensuring that a transportation-efficient semantic representation of the video is preserved.
We build a low-bitrate video understanding benchmark with three downstream tasks on eight datasets, demonstrating the notable superiority of our approach.
arXiv Detail & Related papers (2022-02-06T16:29:15Z) - Low-Fidelity End-to-End Video Encoder Pre-training for Temporal Action
Localization [96.73647162960842]
TAL is a fundamental yet challenging task in video understanding.
Existing TAL methods rely on pre-training a video encoder through action classification supervision.
We introduce a novel low-fidelity end-to-end (LoFi) video encoder pre-training method.
arXiv Detail & Related papers (2021-03-28T22:18:14Z) - Ultra-low bitrate video conferencing using deep image animation [7.263312285502382]
We propose a novel deep learning approach for ultra-low video compression for video conferencing applications.
We employ deep neural networks to encode motion information as keypoint displacement and reconstruct the video signal at the decoder side.
arXiv Detail & Related papers (2020-12-01T09:06:34Z) - Content Adaptive and Error Propagation Aware Deep Video Compression [110.31693187153084]
We propose a content adaptive and error propagation aware video compression system.
Our method employs a joint training strategy by considering the compression performance of multiple consecutive frames instead of a single frame.
Instead of using the hand-crafted coding modes in the traditional compression systems, we design an online encoder updating scheme in our system.
arXiv Detail & Related papers (2020-03-25T09:04:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.