Butterfly: Multiple Reference Frames Feature Propagation Mechanism for
Neural Video Compression
- URL: http://arxiv.org/abs/2303.02959v1
- Date: Mon, 6 Mar 2023 08:19:15 GMT
- Title: Butterfly: Multiple Reference Frames Feature Propagation Mechanism for
Neural Video Compression
- Authors: Feng Wang, Haihang Ruan, Fei Xiong, Jiayu Yang, Litian Li and Ronggang
Wang
- Abstract summary: We present a more reasonable multi-reference frames propagation mechanism for neural video compression.
Our method can significantly outperform the previous state-of-the-art (SOTA)
Our neural dataset can achieve -7.6% save on HEVC Class D when compared with our base single-reference frame model.
- Score: 17.073251238499314
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Using more reference frames can significantly improve the compression
efficiency in neural video compression. However, in low-latency scenarios, most
existing neural video compression frameworks usually use the previous one frame
as reference. Or a few frameworks which use the previous multiple frames as
reference only adopt a simple multi-reference frames propagation mechanism. In
this paper, we present a more reasonable multi-reference frames propagation
mechanism for neural video compression, called butterfly multi-reference frame
propagation mechanism (Butterfly), which allows a more effective feature fusion
of multi-reference frames. By this, we can generate more accurate temporal
context conditional prior for Contextual Coding Module. Besides, when the
number of decoded frames does not meet the required number of reference frames,
we duplicate the nearest reference frame to achieve the requirement, which is
better than duplicating the furthest one. Experiment results show that our
method can significantly outperform the previous state-of-the-art (SOTA), and
our neural codec can achieve -7.6% bitrate save on HEVC Class D dataset when
compares with our base single-reference frame model with the same compression
configuration.
Related papers
- High-Efficiency Neural Video Compression via Hierarchical Predictive Learning [27.41398149573729]
Enhanced Deep Hierarchical Video Compression-DHVC 2.0- introduces superior compression performance and impressive complexity efficiency.
Uses hierarchical predictive coding to transform each video frame into multiscale representations.
Supports transmission-friendly progressive decoding, making it particularly advantageous for networked video applications in the presence of packet loss.
arXiv Detail & Related papers (2024-10-03T15:40:58Z) - IBVC: Interpolation-driven B-frame Video Compression [68.18440522300536]
B-frame video compression aims to adopt bi-directional motion estimation and motion compensation (MEMC) coding for middle frame reconstruction.
Previous learned approaches often directly extend neural P-frame codecs to B-frame relying on bi-directional optical-flow estimation.
We propose a simple yet effective structure called Interpolation-B-frame Video Compression (IBVC) to address these issues.
arXiv Detail & Related papers (2023-09-25T02:45:51Z) - Predictive Coding For Animation-Based Video Compression [13.161311799049978]
We propose a predictive coding scheme which uses image animation as a predictor, and codes the residual with respect to the actual target frame.
Our experiments indicate a significant gain, in excess of 70% compared to the HEVC video standard and over 30% compared to VVC.
arXiv Detail & Related papers (2023-07-09T14:40:54Z) - Advancing Learned Video Compression with In-loop Frame Prediction [177.67218448278143]
In this paper, we propose an Advanced Learned Video Compression (ALVC) approach with the in-loop frame prediction module.
The predicted frame can serve as a better reference than the previously compressed frame, and therefore it benefits the compression performance.
The experiments show the state-of-the-art performance of our ALVC approach in learned video compression.
arXiv Detail & Related papers (2022-11-13T19:53:14Z) - Extending Neural P-frame Codecs for B-frame Coding [15.102346715690755]
Our B-frame solution is based on the existing P-frame methods.
Our results show that using the proposed method with an existing P-frame can lead to 28.5%saving in bit-rate on the UVG dataset.
arXiv Detail & Related papers (2021-03-30T21:25:35Z) - Conditional Entropy Coding for Efficient Video Compression [82.35389813794372]
We propose a very simple and efficient video compression framework that only focuses on modeling the conditional entropy between frames.
We first show that a simple architecture modeling the entropy between the image latent codes is as competitive as other neural video compression works and video codecs.
We then propose a novel internal learning extension on top of this architecture that brings an additional 10% savings without trading off decoding speed.
arXiv Detail & Related papers (2020-08-20T20:01:59Z) - Learning for Video Compression with Recurrent Auto-Encoder and Recurrent
Probability Model [164.7489982837475]
This paper proposes a Recurrent Learned Video Compression (RLVC) approach with the Recurrent Auto-Encoder (RAE) and Recurrent Probability Model ( RPM)
The RAE employs recurrent cells in both the encoder and decoder to exploit the temporal correlation among video frames.
Our approach achieves the state-of-the-art learned video compression performance in terms of both PSNR and MS-SSIM.
arXiv Detail & Related papers (2020-06-24T08:46:33Z) - M-LVC: Multiple Frames Prediction for Learned Video Compression [111.50760486258993]
We propose an end-to-end learned video compression scheme for low-latency scenarios.
In our scheme, the motion vector (MV) field is calculated between the current frame and the previous one.
Experimental results show that the proposed method outperforms the existing learned video compression methods for low-latency mode.
arXiv Detail & Related papers (2020-04-21T20:42:02Z) - Efficient Semantic Video Segmentation with Per-frame Inference [117.97423110566963]
In this work, we process efficient semantic video segmentation in a per-frame fashion during the inference process.
We employ compact models for real-time execution. To narrow the performance gap between compact models and large models, new knowledge distillation methods are designed.
arXiv Detail & Related papers (2020-02-26T12:24:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.