Long-term Video Frame Interpolation via Feature Propagation
- URL: http://arxiv.org/abs/2203.15427v1
- Date: Tue, 29 Mar 2022 10:47:06 GMT
- Title: Long-term Video Frame Interpolation via Feature Propagation
- Authors: Dawit Mureja Argaw and In So Kweon
- Abstract summary: Video frame (VFI) works generally predict intermediate frame(s) by first estimating the motion between inputs and then warping the inputs to the target time with the estimated motion.
This approach is not optimal when the temporal distance between the input sequence increases.
We propose a propagation network (PNet) by extending the classic feature-level forecasting with a novel motion-to-feature approach.
- Score: 95.18170372022703
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Video frame interpolation (VFI) works generally predict intermediate frame(s)
by first estimating the motion between inputs and then warping the inputs to
the target time with the estimated motion. This approach, however, is not
optimal when the temporal distance between the input sequence increases as
existing motion estimation modules cannot effectively handle large motions.
Hence, VFI works perform well for small frame gaps and perform poorly as the
frame gap increases. In this work, we propose a novel framework to address this
problem. We argue that when there is a large gap between inputs, instead of
estimating imprecise motion that will eventually lead to inaccurate
interpolation, we can safely propagate from one side of the input up to a
reliable time frame using the other input as a reference. Then, the rest of the
intermediate frames can be interpolated using standard approaches as the
temporal gap is now narrowed. To this end, we propose a propagation network
(PNet) by extending the classic feature-level forecasting with a novel
motion-to-feature approach. To be thorough, we adopt a simple interpolation
model along with PNet as our full model and design a simple procedure to train
the full model in an end-to-end manner. Experimental results on several
benchmark datasets confirm the effectiveness of our method for long-term VFI
compared to state-of-the-art approaches.
Related papers
- Motion-aware Latent Diffusion Models for Video Frame Interpolation [51.78737270917301]
Motion estimation between neighboring frames plays a crucial role in avoiding motion ambiguity.
We propose a novel diffusion framework, motion-aware latent diffusion models (MADiff)
Our method achieves state-of-the-art performance significantly outperforming existing approaches.
arXiv Detail & Related papers (2024-04-21T05:09:56Z) - Motion-Aware Video Frame Interpolation [49.49668436390514]
We introduce a Motion-Aware Video Frame Interpolation (MA-VFI) network, which directly estimates intermediate optical flow from consecutive frames.
It not only extracts global semantic relationships and spatial details from input frames with different receptive fields, but also effectively reduces the required computational cost and complexity.
arXiv Detail & Related papers (2024-02-05T11:00:14Z) - A Multi-In-Single-Out Network for Video Frame Interpolation without
Optical Flow [14.877766449009119]
deep learning-based video frame (VFI) methods have predominantly focused on estimating motion between two input frames.
We propose a multi-in-single-out (MISO) based VFI method that does not rely on motion vector estimation.
We introduce a novel motion perceptual loss that enables MISO-VFI to better capture the vectors-temporal within the video frames.
arXiv Detail & Related papers (2023-11-20T08:29:55Z) - Video Frame Interpolation with Densely Queried Bilateral Correlation [52.823751291070906]
Video Frame Interpolation (VFI) aims to synthesize non-existent intermediate frames between existent frames.
Flow-based VFI algorithms estimate intermediate motion fields to warp the existent frames.
We propose Densely Queried Bilateral Correlation (DQBC) that gets rid of the receptive field dependency problem.
arXiv Detail & Related papers (2023-04-26T14:45:09Z) - Enhanced Bi-directional Motion Estimation for Video Frame Interpolation [0.05541644538483946]
We present a novel yet effective algorithm for motion-based video frame estimation.
Our method achieves excellent performance on a broad range of video frame benchmarks.
arXiv Detail & Related papers (2022-06-17T06:08:43Z) - FLAVR: Flow-Agnostic Video Representations for Fast Frame Interpolation [97.99012124785177]
FLAVR is a flexible and efficient architecture that uses 3D space-time convolutions to enable end-to-end learning and inference for video framesupervised.
We demonstrate that FLAVR can serve as a useful self- pretext task for action recognition, optical flow estimation, and motion magnification.
arXiv Detail & Related papers (2020-12-15T18:59:30Z) - All at Once: Temporally Adaptive Multi-Frame Interpolation with Advanced
Motion Modeling [52.425236515695914]
State-of-the-art methods are iterative solutions interpolating one frame at the time.
This work introduces a true multi-frame interpolator.
It utilizes a pyramidal style network in the temporal domain to complete the multi-frame task in one-shot.
arXiv Detail & Related papers (2020-07-23T02:34:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.