Transcoded Video Restoration by Temporal Spatial Auxiliary Network
- URL: http://arxiv.org/abs/2112.07948v1
- Date: Wed, 15 Dec 2021 08:10:23 GMT
- Title: Transcoded Video Restoration by Temporal Spatial Auxiliary Network
- Authors: Li Xu, Gang He, Jinjia Zhou, Jie Lei, Weiying Xie, Yunsong Li, Yu-Wing
Tai
- Abstract summary: We propose a new method, temporal spatial auxiliary network (TSAN), for transcoded video restoration.
The experimental results demonstrate that the performance of the proposed method is superior to that of the previous techniques.
- Score: 64.63157339057912
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In most video platforms, such as Youtube, and TikTok, the played videos
usually have undergone multiple video encodings such as hardware encoding by
recording devices, software encoding by video editing apps, and single/multiple
video transcoding by video application servers. Previous works in compressed
video restoration typically assume the compression artifacts are caused by
one-time encoding. Thus, the derived solution usually does not work very well
in practice. In this paper, we propose a new method, temporal spatial auxiliary
network (TSAN), for transcoded video restoration. Our method considers the
unique traits between video encoding and transcoding, and we consider the
initial shallow encoded videos as the intermediate labels to assist the network
to conduct self-supervised attention training. In addition, we employ adjacent
multi-frame information and propose the temporal deformable alignment and
pyramidal spatial fusion for transcoded video restoration. The experimental
results demonstrate that the performance of the proposed method is superior to
that of the previous techniques. The code is available at
https://github.com/icecherylXuli/TSAN.
Related papers
- Fast Encoding and Decoding for Implicit Video Representation [88.43612845776265]
We introduce NeRV-Enc, a transformer-based hyper-network for fast encoding; and NeRV-Dec, a parallel decoder for efficient video loading.
NeRV-Enc achieves an impressive speed-up of $mathbf104times$ by eliminating gradient-based optimization.
NeRV-Dec simplifies video decoding, outperforming conventional codecs with a loading speed $mathbf11times$ faster.
arXiv Detail & Related papers (2024-09-28T18:21:52Z) - Predictive Coding For Animation-Based Video Compression [13.161311799049978]
We propose a predictive coding scheme which uses image animation as a predictor, and codes the residual with respect to the actual target frame.
Our experiments indicate a significant gain, in excess of 70% compared to the HEVC video standard and over 30% compared to VVC.
arXiv Detail & Related papers (2023-07-09T14:40:54Z) - Video Compression with Arbitrary Rescaling Network [8.489428003916622]
We propose a rate-guided arbitrary rescaling network (RARN) for video resizing before encoding.
The lightweight RARN structure can process FHD (1080p) content at real-time speed (91 FPS) and obtain a considerable rate reduction.
arXiv Detail & Related papers (2023-06-07T07:15:18Z) - Lightweight Hybrid Video Compression Framework Using Reference-Guided
Restoration Network [12.033330902093152]
We propose a new lightweight hybrid video consisting of a conventional video(HEVC / VVC), a lossless image, and our new restoration network.
Our method achieves comparable performance to top-tier methods, even when applied to HEVC.
arXiv Detail & Related papers (2023-03-21T04:42:44Z) - Compressed Vision for Efficient Video Understanding [83.97689018324732]
We propose a framework enabling research on hour-long videos with the same hardware that can now process second-long videos.
We replace standard video compression, e.g. JPEG, with neural compression and show that we can directly feed compressed videos as inputs to regular video networks.
arXiv Detail & Related papers (2022-10-06T15:35:49Z) - A Coding Framework and Benchmark towards Low-Bitrate Video Understanding [63.05385140193666]
We propose a traditional-neural mixed coding framework that takes advantage of both traditional codecs and neural networks (NNs)
The framework is optimized by ensuring that a transportation-efficient semantic representation of the video is preserved.
We build a low-bitrate video understanding benchmark with three downstream tasks on eight datasets, demonstrating the notable superiority of our approach.
arXiv Detail & Related papers (2022-02-06T16:29:15Z) - Deep Contextual Video Compression [20.301569390401102]
We propose a deep contextual video compression framework to enable a paradigm shift from predictive coding to conditional coding.
Our method can significantly outperform the previous state-of-theart (SOTA) deep video compression methods.
arXiv Detail & Related papers (2021-09-30T12:14:24Z) - VPN: Video Provenance Network for Robust Content Attribution [72.12494245048504]
We present VPN - a content attribution method for recovering provenance information from videos shared online.
We learn a robust search embedding for matching such video, using full-length or truncated video queries.
Once matched against a trusted database of video clips, associated information on the provenance of the clip is presented to the user.
arXiv Detail & Related papers (2021-09-21T09:07:05Z) - Content Adaptive and Error Propagation Aware Deep Video Compression [110.31693187153084]
We propose a content adaptive and error propagation aware video compression system.
Our method employs a joint training strategy by considering the compression performance of multiple consecutive frames instead of a single frame.
Instead of using the hand-crafted coding modes in the traditional compression systems, we design an online encoder updating scheme in our system.
arXiv Detail & Related papers (2020-03-25T09:04:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.