End-To-End Trainable Video Super-Resolution Based on a New Mechanism for
Implicit Motion Estimation and Compensation
- URL: http://arxiv.org/abs/2001.01162v1
- Date: Sun, 5 Jan 2020 03:47:24 GMT
- Title: End-To-End Trainable Video Super-Resolution Based on a New Mechanism for
Implicit Motion Estimation and Compensation
- Authors: Xiaohong Liu, Lingshi Kong, Yang Zhou, Jiying Zhao, Jun Chen
- Abstract summary: Video super-resolution aims at generating a high-resolution video from its low-resolution counterpart.
We propose a novel dynamic local filter network to perform implicit motion estimation and compensation.
We also propose a global refinement network based on ResBlock and autoencoder structures.
- Score: 19.67999205691758
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video super-resolution aims at generating a high-resolution video from its
low-resolution counterpart. With the rapid rise of deep learning, many recently
proposed video super-resolution methods use convolutional neural networks in
conjunction with explicit motion compensation to capitalize on statistical
dependencies within and across low-resolution frames. Two common issues of such
methods are noteworthy. Firstly, the quality of the final reconstructed HR
video is often very sensitive to the accuracy of motion estimation. Secondly,
the warp grid needed for motion compensation, which is specified by the two
flow maps delineating pixel displacements in horizontal and vertical
directions, tends to introduce additional errors and jeopardize the temporal
consistency across video frames. To address these issues, we propose a novel
dynamic local filter network to perform implicit motion estimation and
compensation by employing, via locally connected layers, sample-specific and
position-specific dynamic local filters that are tailored to the target pixels.
We also propose a global refinement network based on ResBlock and autoencoder
structures to exploit non-local correlations and enhance the spatial
consistency of super-resolved frames. The experimental results demonstrate that
the proposed method outperforms the state-of-the-art, and validate its strength
in terms of local transformation handling, temporal consistency as well as edge
sharpness.
Related papers
- Residual Learning and Filtering Networks for End-to-End Lossless Video Compression [3.0770091134672586]
Existing learning-based video compression methods face challenges related to inaccurate motion estimates and inadequate motion compensation structures.
This work presents an end-to-end video compression method that incorporates several key operations.
The proposed approach tackles the challenges of accurate motion estimation and motion compensation in video compression.
arXiv Detail & Related papers (2025-03-11T18:51:36Z) - CoordFlow: Coordinate Flow for Pixel-wise Neural Video Representation [11.364753833652182]
Implicit Neural Representation (INR) is a promising alternative to traditional transform-based methodologies.
We introduce CoordFlow, a novel pixel-wise INR for video compression.
It yields state-of-the-art results compared to other pixel-wise INRs and on-par performance compared to leading frame-wise techniques.
arXiv Detail & Related papers (2025-01-01T22:58:06Z) - Spatio-Temporal Distortion Aware Omnidirectional Video Super-Resolution [26.166579083377556]
Video super-resolution (VSR) methods are proposed to enhance the resolution of videos, but ODV projection distortions are not well addressed directly applying such methods.
We propose a novel Spatio-Temporal Distortion Aware Network (STDAN) to achieve better super-resolution reconstruction quality.
arXiv Detail & Related papers (2024-10-15T11:17:19Z) - Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - Learning Spatial Adaptation and Temporal Coherence in Diffusion Models for Video Super-Resolution [151.1255837803585]
We propose a novel approach, pursuing Spatial Adaptation and Temporal Coherence (SATeCo) for video super-resolution.
SATeCo pivots on learning spatial-temporal guidance from low-resolution videos to calibrate both latent-space high-resolution video denoising and pixel-space video reconstruction.
Experiments conducted on the REDS4 and Vid4 datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-03-25T17:59:26Z) - Differentiable Resolution Compression and Alignment for Efficient Video
Classification and Retrieval [16.497758750494537]
We propose an efficient video representation network with Differentiable Resolution Compression and Alignment mechanism.
We leverage a Differentiable Context-aware Compression Module to encode the saliency and non-saliency frame features.
We introduce a new Resolution-Align Transformer Layer to capture global temporal correlations among frame features with different resolutions.
arXiv Detail & Related papers (2023-09-15T05:31:53Z) - Towards Interpretable Video Super-Resolution via Alternating
Optimization [115.85296325037565]
We study a practical space-time video super-resolution (STVSR) problem which aims at generating a high-framerate high-resolution sharp video from a low-framerate blurry video.
We propose an interpretable STVSR framework by leveraging both model-based and learning-based methods.
arXiv Detail & Related papers (2022-07-21T21:34:05Z) - Fractional Motion Estimation for Point Cloud Compression [34.6429671080134]
We show that motion compensation can benefit from the higher resolution reference and more accurate displacements provided by fractional precision.
The proposed scheme can be combined with and add sizeable gains to state-of-the-art systems that use transforms such as Region Adaptive Graph Fourier Transform and Region Adaptive Haar Transform.
arXiv Detail & Related papers (2022-02-01T01:00:28Z) - Optical-Flow-Reuse-Based Bidirectional Recurrent Network for Space-Time
Video Super-Resolution [52.899234731501075]
Space-time video super-resolution (ST-VSR) simultaneously increases the spatial resolution and frame rate for a given video.
Existing methods typically suffer from difficulties in how to efficiently leverage information from a large range of neighboring frames.
We propose a coarse-to-fine bidirectional recurrent neural network instead of using ConvLSTM to leverage knowledge between adjacent frames.
arXiv Detail & Related papers (2021-10-13T15:21:30Z) - Zooming Slow-Mo: Fast and Accurate One-Stage Space-Time Video
Super-Resolution [95.26202278535543]
A simple solution is to split it into two sub-tasks: video frame (VFI) and video super-resolution (VSR)
temporalsynthesis and spatial super-resolution are intra-related in this task.
We propose a one-stage space-time video super-resolution framework, which directly synthesizes an HR slow-motion video from an LFR, LR video.
arXiv Detail & Related papers (2020-02-26T16:59:48Z) - Video Face Super-Resolution with Motion-Adaptive Feedback Cell [90.73821618795512]
Video super-resolution (VSR) methods have recently achieved a remarkable success due to the development of deep convolutional neural networks (CNN)
In this paper, we propose a Motion-Adaptive Feedback Cell (MAFC), a simple but effective block, which can efficiently capture the motion compensation and feed it back to the network in an adaptive way.
arXiv Detail & Related papers (2020-02-15T13:14:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.