EFENet: Reference-based Video Super-Resolution with Enhanced Flow
Estimation
- URL: http://arxiv.org/abs/2110.07797v1
- Date: Fri, 15 Oct 2021 01:36:30 GMT
- Title: EFENet: Reference-based Video Super-Resolution with Enhanced Flow
Estimation
- Authors: Yaping Zhao, Mengqi Ji, Ruqi Huang, Bin Wang, Shengjin Wang
- Abstract summary: We propose EFENet to exploit simultaneously the visual cues contained in the HR reference and the temporal information contained in the LR sequence.
We provide evaluations to validate the strengths of our approach, and to demonstrate that the proposed framework outperforms the state-of-the-art methods.
- Score: 33.170496636269114
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we consider the problem of reference-based video
super-resolution(RefVSR), i.e., how to utilize a high-resolution (HR) reference
frame to super-resolve a low-resolution (LR) video sequence. The existing
approaches to RefVSR essentially attempt to align the reference and the input
sequence, in the presence of resolution gap and long temporal range. However,
they either ignore temporal structure within the input sequence, or suffer
accumulative alignment errors. To address these issues, we propose EFENet to
exploit simultaneously the visual cues contained in the HR reference and the
temporal information contained in the LR sequence. EFENet first globally
estimates cross-scale flow between the reference and each LR frame. Then our
novel flow refinement module of EFENet refines the flow regarding the furthest
frame using all the estimated flows, which leverages the global temporal
information within the sequence and therefore effectively reduces the alignment
errors. We provide comprehensive evaluations to validate the strengths of our
approach, and to demonstrate that the proposed framework outperforms the
state-of-the-art methods. Code is available at
https://github.com/IndigoPurple/EFENet.
Related papers
- Event-Enhanced Blurry Video Super-Resolution [52.894824081586776]
We tackle the task of blurry video super-resolution (BVSR), aiming to generate high-resolution (HR) videos from low-resolution (LR) and blurry inputs.
Current BVSR methods often fail to restore sharp details at high resolutions, resulting in noticeable artifacts and jitter.
We introduce event signals into BVSR and propose a novel event-enhanced network, Ev-DeVSR.
arXiv Detail & Related papers (2025-04-17T15:55:41Z) - Sparse Global Matching for Video Frame Interpolation with Large Motion [20.49084881829404]
Large motion poses a critical challenge in Video Frame Interpolation (VFI) task.
Existing methods are often constrained by limited receptive fields, resulting in sub-optimal performance when handling scenarios with large motion.
We introduce a new pipeline for VFI, which can effectively integrate global-level information to alleviate issues associated with large motion.
arXiv Detail & Related papers (2024-04-10T11:06:29Z) - StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video
Sequences [31.210626775505407]
Occlusions between consecutive frames have long posed a significant challenge in optical flow estimation.
We present a Streamlined In-batch Multi-frame (SIM) pipeline tailored to video input, attaining a similar level of time efficiency to two-frame networks.
StreamFlow not only excels in terms of performance on challenging KITTI and Sintel datasets, with particular improvement in occluded areas.
arXiv Detail & Related papers (2023-11-28T07:53:51Z) - Local-Global Temporal Difference Learning for Satellite Video
Super-Resolution [55.69322525367221]
We propose to exploit the well-defined temporal difference for efficient and effective temporal compensation.
To fully utilize the local and global temporal information within frames, we systematically modeled the short-term and long-term temporal discrepancies.
Rigorous objective and subjective evaluations conducted across five mainstream video satellites demonstrate that our method performs favorably against state-of-the-art approaches.
arXiv Detail & Related papers (2023-04-10T07:04:40Z) - Continuous Space-Time Video Super-Resolution Utilizing Long-Range
Temporal Information [48.20843501171717]
We propose a continuous ST-VSR (CSTVSR) method that can convert the given video to any frame rate and spatial resolution.
We show that the proposed algorithm has good flexibility and achieves better performance on various datasets.
arXiv Detail & Related papers (2023-02-26T08:02:39Z) - Temporal Consistency Learning of inter-frames for Video Super-Resolution [38.26035126565062]
Video super-resolution (VSR) is a task that aims to reconstruct high-resolution (HR) frames from the low-resolution (LR) reference frame and multiple neighboring frames.
Existing methods generally explore information propagation and frame alignment to improve the performance of VSR.
We propose a Temporal Consistency learning Network (TCNet) for VSR in an end-to-end manner, to enhance the consistency of the reconstructed videos.
arXiv Detail & Related papers (2022-11-03T08:23:57Z) - Fast Online Video Super-Resolution with Deformable Attention Pyramid [172.16491820970646]
Video super-resolution (VSR) has many applications that pose strict causal, real-time, and latency constraints, including video streaming and TV.
We propose a recurrent VSR architecture based on a deformable attention pyramid (DAP)
arXiv Detail & Related papers (2022-02-03T17:49:04Z) - Optical-Flow-Reuse-Based Bidirectional Recurrent Network for Space-Time
Video Super-Resolution [52.899234731501075]
Space-time video super-resolution (ST-VSR) simultaneously increases the spatial resolution and frame rate for a given video.
Existing methods typically suffer from difficulties in how to efficiently leverage information from a large range of neighboring frames.
We propose a coarse-to-fine bidirectional recurrent neural network instead of using ConvLSTM to leverage knowledge between adjacent frames.
arXiv Detail & Related papers (2021-10-13T15:21:30Z) - FDAN: Flow-guided Deformable Alignment Network for Video
Super-Resolution [12.844337773258678]
Flow-guided Deformable Module (FDM) is proposed to integrate optical flow into deformable convolution.
FDAN reaches the state-of-the-art performance on two benchmark datasets.
arXiv Detail & Related papers (2021-05-12T13:18:36Z) - Video Super-Resolution with Recurrent Structure-Detail Network [120.1149614834813]
Most video super-resolution methods super-resolve a single reference frame with the help of neighboring frames in a temporal sliding window.
We propose a novel recurrent video super-resolution method which is both effective and efficient in exploiting previous frames to super-resolve the current frame.
arXiv Detail & Related papers (2020-08-02T11:01:19Z) - Video Face Super-Resolution with Motion-Adaptive Feedback Cell [90.73821618795512]
Video super-resolution (VSR) methods have recently achieved a remarkable success due to the development of deep convolutional neural networks (CNN)
In this paper, we propose a Motion-Adaptive Feedback Cell (MAFC), a simple but effective block, which can efficiently capture the motion compensation and feed it back to the network in an adaptive way.
arXiv Detail & Related papers (2020-02-15T13:14:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.