Video Frame Interpolation Based on Deformable Kernel Region
- URL: http://arxiv.org/abs/2204.11396v1
- Date: Mon, 25 Apr 2022 02:03:04 GMT
- Title: Video Frame Interpolation Based on Deformable Kernel Region
- Authors: Haoyue Tian, Pan Gao, Xiaojiang Peng
- Abstract summary: We propose a deformable convolution for video, which can break the fixed grid restrictions on the kernel region.
Experiments are conducted on four datasets to demonstrate the superior performance of the proposed model.
- Score: 18.55904569126297
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video frame interpolation task has recently become more and more prevalent in
the computer vision field. At present, a number of researches based on deep
learning have achieved great success. Most of them are either based on optical
flow information, or interpolation kernel, or a combination of these two
methods. However, these methods have ignored that there are grid restrictions
on the position of kernel region during synthesizing each target pixel. These
limitations result in that they cannot well adapt to the irregularity of object
shape and uncertainty of motion, which may lead to irrelevant reference pixels
used for interpolation. In order to solve this problem, we revisit the
deformable convolution for video interpolation, which can break the fixed grid
restrictions on the kernel region, making the distribution of reference points
more suitable for the shape of the object, and thus warp a more accurate
interpolation frame. Experiments are conducted on four datasets to demonstrate
the superior performance of the proposed model in comparison to the
state-of-the-art alternatives.
Related papers
- ViBiDSampler: Enhancing Video Interpolation Using Bidirectional Diffusion Sampler [53.98558445900626]
Current image-to-video diffusion models, while powerful in generating videos from a single frame, need adaptation for two-frame conditioned generation.
We introduce a novel, bidirectional sampling strategy to address these off-manifold issues without requiring extensive re-noising or fine-tuning.
Our method employs sequential sampling along both forward and backward paths, conditioned on the start and end frames, respectively, ensuring more coherent and on-manifold generation of intermediate frames.
arXiv Detail & Related papers (2024-10-08T03:01:54Z) - Neighbor Correspondence Matching for Flow-based Video Frame Synthesis [90.14161060260012]
We introduce a neighbor correspondence matching (NCM) algorithm for flow-based frame synthesis.
NCM is performed in a current-frame-agnostic fashion to establish multi-scale correspondences in the spatial-temporal neighborhoods of each pixel.
coarse-scale module is designed to leverage neighbor correspondences to capture large motion, while the fine-scale module is more efficient to speed up the estimation process.
arXiv Detail & Related papers (2022-07-14T09:17:00Z) - Deep Recurrent Neural Network with Multi-scale Bi-directional
Propagation for Video Deblurring [36.94523101375519]
We propose a deep Recurrent Neural Network with Multi-scale Bi-directional Propagation (RNN-MBP) to propagate and gather information from unaligned neighboring frames for better video deblurring.
To better evaluate the proposed algorithm and existing state-of-the-art methods on real-world blurry scenes, we also create a Real-World Blurry Video dataset.
The proposed algorithm performs favorably against the state-of-the-art methods on three typical benchmarks.
arXiv Detail & Related papers (2021-12-09T11:02:56Z) - Video Frame Interpolation Transformer [86.20646863821908]
We propose a Transformer-based video framework that allows content-aware aggregation weights and considers long-range dependencies with the self-attention operations.
To avoid the high computational cost of global self-attention, we introduce the concept of local attention into video.
In addition, we develop a multi-scale frame scheme to fully realize the potential of Transformers.
arXiv Detail & Related papers (2021-11-27T05:35:10Z) - TimeLens: Event-based Video Frame Interpolation [54.28139783383213]
We introduce Time Lens, a novel indicates equal contribution method that leverages the advantages of both synthesis-based and flow-based approaches.
We show an up to 5.21 dB improvement in terms of PSNR over state-of-the-art frame-based and event-based methods.
arXiv Detail & Related papers (2021-06-14T10:33:47Z) - ARVo: Learning All-Range Volumetric Correspondence for Video Deblurring [92.40655035360729]
Video deblurring models exploit consecutive frames to remove blurs from camera shakes and object motions.
We propose a novel implicit method to learn spatial correspondence among blurry frames in the feature space.
Our proposed method is evaluated on the widely-adopted DVD dataset, along with a newly collected High-Frame-Rate (1000 fps) dataset for Video Deblurring.
arXiv Detail & Related papers (2021-03-07T04:33:13Z) - Video Frame Interpolation via Generalized Deformable Convolution [18.357839820102683]
Video frame aims at synthesizing intermediate frames from nearby source frames while maintaining spatial and temporal consistencies.
Existing deeplearning-based video frame methods can be divided into two categories: flow-based methods and kernel-based methods.
A novel mechanism named generalized deformable convolution is proposed, which can effectively learn motion in a data-driven manner and freely select sampling points in space-time.
arXiv Detail & Related papers (2020-08-24T20:00:39Z) - Multiple Video Frame Interpolation via Enhanced Deformable Separable
Convolution [67.83074893311218]
Kernel-based methods predict pixels with a single convolution process that convolves source frames with spatially adaptive local kernels.
We propose enhanced deformable separable convolution (EDSC) to estimate not only adaptive kernels, but also offsets, masks and biases.
We show that our method performs favorably against the state-of-the-art methods across a broad range of datasets.
arXiv Detail & Related papers (2020-06-15T01:10:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.