Dynamic Video Frame Interpolation with integrated Difficulty
Pre-Assessment
- URL: http://arxiv.org/abs/2304.12664v1
- Date: Tue, 25 Apr 2023 09:11:20 GMT
- Title: Dynamic Video Frame Interpolation with integrated Difficulty
Pre-Assessment
- Authors: Ban Chen, Xin Jin, Youxin Chen, Longhai Wu, Jie Chen, Jayoon Koo,
Cheul-hee Hahm
- Abstract summary: Video frame(VFI) models still struggle to achieve a good trade-off between accuracy and efficiency.
We present an integrated pipeline which combines difficulty assessment with video frame dataset.
Our proposed pipeline can improve the accuracy-efficiency trade-off for VFI.
- Score: 10.248729137820442
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video frame interpolation(VFI) has witnessed great progress in recent years.
While existing VFI models still struggle to achieve a good trade-off between
accuracy and efficiency: fast models often have inferior accuracy; accurate
models typically run slowly. However, easy samples with small motion or clear
texture can achieve competitive results with simple models and do not require
heavy computation. In this paper, we present an integrated pipeline which
combines difficulty assessment with video frame interpolation. Specifically, it
firstly leverages a pre-assessment model to measure the interpolation
difficulty level of input frames, and then dynamically selects an appropriate
VFI model to generate interpolation results. Furthermore, a large-scale VFI
difficulty assessment dataset is collected and annotated to train our
pre-assessment model. Extensive experiments show that easy samples pass through
fast models while difficult samples inference with heavy models, and our
proposed pipeline can improve the accuracy-efficiency trade-off for VFI.
Related papers
- Motion-aware Latent Diffusion Models for Video Frame Interpolation [51.78737270917301]
Motion estimation between neighboring frames plays a crucial role in avoiding motion ambiguity.
We propose a novel diffusion framework, motion-aware latent diffusion models (MADiff)
Our method achieves state-of-the-art performance significantly outperforming existing approaches.
arXiv Detail & Related papers (2024-04-21T05:09:56Z) - Video Interpolation with Diffusion Models [54.06746595879689]
We present VIDIM, a generative model for video, which creates short videos given a start and end frame.
VIDIM uses cascaded diffusion models to first generate the target video at low resolution, and then generate the high-resolution video conditioned on the low-resolution generated video.
arXiv Detail & Related papers (2024-04-01T15:59:32Z) - FrameNeRF: A Simple and Efficient Framework for Few-shot Novel View
Synthesis [25.356376402671536]
FrameNeRF is designed to apply off-the-shelf fast high-fidelity NeRF models with fast training speed and high rendering quality for few-shot novel view synthesis tasks.
arXiv Detail & Related papers (2024-02-22T14:41:02Z) - Empirical Analysis of Efficient Fine-Tuning Methods for Large
Pre-Trained Language Models [4.096453902709292]
BitFit and adapter modules are compared to standard full model fine-tuning.
The BitFit approach matches full fine-tuning performance across varying amounts of training data.
adapter modules exhibit high variability, with inconsistent gains over default models.
arXiv Detail & Related papers (2024-01-08T17:44:43Z) - Boost Video Frame Interpolation via Motion Adaptation [73.42573856943923]
Video frame (VFI) is a challenging task that aims to generate intermediate frames between two consecutive frames in a video.
Existing learning-based VFI methods have achieved great success, but they still suffer from limited generalization ability.
We propose a novel optimization-based VFI method that can adapt to unseen motions at test time.
arXiv Detail & Related papers (2023-06-24T10:44:02Z) - H-VFI: Hierarchical Frame Interpolation for Videos with Large Motions [63.23985601478339]
We propose a simple yet effective solution, H-VFI, to deal with large motions in video frame.
H-VFI contributes a hierarchical video transformer to learn a deformable kernel in a coarse-to-fine strategy.
The advantage of such a progressive approximation is that the large motion frame problem can be predicted into several relatively simpler sub-tasks.
arXiv Detail & Related papers (2022-11-21T09:49:23Z) - Long-term Video Frame Interpolation via Feature Propagation [95.18170372022703]
Video frame (VFI) works generally predict intermediate frame(s) by first estimating the motion between inputs and then warping the inputs to the target time with the estimated motion.
This approach is not optimal when the temporal distance between the input sequence increases.
We propose a propagation network (PNet) by extending the classic feature-level forecasting with a novel motion-to-feature approach.
arXiv Detail & Related papers (2022-03-29T10:47:06Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.