IFRNet: Intermediate Feature Refine Network for Efficient Frame
Interpolation
- URL: http://arxiv.org/abs/2205.14620v1
- Date: Sun, 29 May 2022 10:18:18 GMT
- Title: IFRNet: Intermediate Feature Refine Network for Efficient Frame
Interpolation
- Authors: Lingtong Kong, Boyuan Jiang, Donghao Luo, Wenqing Chu, Xiaoming Huang,
Ying Tai, Chengjie Wang, Jie Yang
- Abstract summary: We devise an efficient encoder-decoder based network, termed IFRNet, for fast intermediate frame synthesizing.
Experiments on various benchmarks demonstrate the excellent performance and fast inference speed of proposed approaches.
- Score: 44.04110765492441
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prevailing video frame interpolation algorithms, that generate the
intermediate frames from consecutive inputs, typically rely on complex model
architectures with heavy parameters or large delay, hindering them from diverse
real-time applications. In this work, we devise an efficient encoder-decoder
based network, termed IFRNet, for fast intermediate frame synthesizing. It
first extracts pyramid features from given inputs, and then refines the
bilateral intermediate flow fields together with a powerful intermediate
feature until generating the desired output. The gradually refined intermediate
feature can not only facilitate intermediate flow estimation, but also
compensate for contextual details, making IFRNet do not need additional
synthesis or refinement module. To fully release its potential, we further
propose a novel task-oriented optical flow distillation loss to focus on
learning the useful teacher knowledge towards frame synthesizing. Meanwhile, a
new geometry consistency regularization term is imposed on the gradually
refined intermediate features to keep better structure layout. Experiments on
various benchmarks demonstrate the excellent performance and fast inference
speed of proposed approaches. Code is available at
https://github.com/ltkong218/IFRNet.
Related papers
- Motion-Aware Video Frame Interpolation [49.49668436390514]
We introduce a Motion-Aware Video Frame Interpolation (MA-VFI) network, which directly estimates intermediate optical flow from consecutive frames.
It not only extracts global semantic relationships and spatial details from input frames with different receptive fields, but also effectively reduces the required computational cost and complexity.
arXiv Detail & Related papers (2024-02-05T11:00:14Z) - StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video
Sequences [31.210626775505407]
Occlusions between consecutive frames have long posed a significant challenge in optical flow estimation.
We present a Streamlined In-batch Multi-frame (SIM) pipeline tailored to video input, attaining a similar level of time efficiency to two-frame networks.
StreamFlow not only excels in terms of performance on challenging KITTI and Sintel datasets, with particular improvement in occluded areas.
arXiv Detail & Related papers (2023-11-28T07:53:51Z) - Dynamic Frame Interpolation in Wavelet Domain [57.25341639095404]
Video frame is an important low-level computation vision task, which can increase frame rate for more fluent visual experience.
Existing methods have achieved great success by employing advanced motion models and synthesis networks.
WaveletVFI can reduce computation up to 40% while maintaining similar accuracy, making it perform more efficiently against other state-of-the-arts.
arXiv Detail & Related papers (2023-09-07T06:41:15Z) - Progressive Motion Context Refine Network for Efficient Video Frame
Interpolation [10.369068266836154]
Flow-based frame methods have achieved great success by first modeling optical flow between target and input frames, and then building synthesis network for target frame generation.
We propose a novel Progressive Motion Context Refine Network (PMCRNet) to predict motion fields and image context jointly for higher efficiency.
Experiments on multiple benchmarks show that proposed approaches not only achieve favorable and quantitative results but also reduces model size and running time significantly.
arXiv Detail & Related papers (2022-11-11T06:29:03Z) - A Unified Pyramid Recurrent Network for Video Frame Interpolation [10.859715542859773]
We present UPR-Net, a novel Unified Pyramid Recurrent Network for frame synthesis.
We show that our iterative synthesis strategy can significantly improve the robustness of the frame on large motion cases.
arXiv Detail & Related papers (2022-11-07T11:12:31Z) - Meta-Interpolation: Time-Arbitrary Frame Interpolation via Dual
Meta-Learning [65.85319901760478]
We consider processing different time-steps with adaptively generated convolutional kernels in a unified way with the help of meta-learning.
We develop a dual meta-learned frame framework to synthesize intermediate frames with the guidance of context information and optical flow.
arXiv Detail & Related papers (2022-07-27T17:36:23Z) - Long-term Video Frame Interpolation via Feature Propagation [95.18170372022703]
Video frame (VFI) works generally predict intermediate frame(s) by first estimating the motion between inputs and then warping the inputs to the target time with the estimated motion.
This approach is not optimal when the temporal distance between the input sequence increases.
We propose a propagation network (PNet) by extending the classic feature-level forecasting with a novel motion-to-feature approach.
arXiv Detail & Related papers (2022-03-29T10:47:06Z) - FLAVR: Flow-Agnostic Video Representations for Fast Frame Interpolation [97.99012124785177]
FLAVR is a flexible and efficient architecture that uses 3D space-time convolutions to enable end-to-end learning and inference for video framesupervised.
We demonstrate that FLAVR can serve as a useful self- pretext task for action recognition, optical flow estimation, and motion magnification.
arXiv Detail & Related papers (2020-12-15T18:59:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.