Video Super-resolution with Temporal Group Attention
- URL: http://arxiv.org/abs/2007.10595v1
- Date: Tue, 21 Jul 2020 04:54:30 GMT
- Title: Video Super-resolution with Temporal Group Attention
- Authors: Takashi Isobe, Songjiang Li, Xu Jia, Shanxin Yuan, Gregory Slabaugh,
Chunjing Xu, Ya-Li Li, Shengjin Wang, Qi Tian
- Abstract summary: We propose a novel method that can effectively incorporate temporal information in a hierarchical way.
The input sequence is divided into several groups, with each one corresponding to a kind of frame rate.
It achieves favorable performance against state-of-the-art methods on several benchmark datasets.
- Score: 127.21615040695941
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video super-resolution, which aims at producing a high-resolution video from
its corresponding low-resolution version, has recently drawn increasing
attention. In this work, we propose a novel method that can effectively
incorporate temporal information in a hierarchical way. The input sequence is
divided into several groups, with each one corresponding to a kind of frame
rate. These groups provide complementary information to recover missing details
in the reference frame, which is further integrated with an attention module
and a deep intra-group fusion module. In addition, a fast spatial alignment is
proposed to handle videos with large motion. Extensive results demonstrate the
capability of the proposed model in handling videos with various motion. It
achieves favorable performance against state-of-the-art methods on several
benchmark datasets.
Related papers
- Conditional Modeling Based Automatic Video Summarization [70.96973928590958]
The aim of video summarization is to shorten videos automatically while retaining the key information necessary to convey the overall story.
Video summarization methods rely on visual factors, such as visual consecutiveness and diversity, which may not be sufficient to fully understand the content of the video.
A new approach to video summarization is proposed based on insights gained from how humans create ground truth video summaries.
arXiv Detail & Related papers (2023-11-20T20:24:45Z) - A Lightweight Recurrent Grouping Attention Network for Video
Super-Resolution [0.0]
We propose a lightweight lightweight grouping attention network to reduce the stress on the device.
The parameters of this model are only 0.878M, which is much lower than the current mainstream model for studying video super-resolution.
Experiments demonstrate that our model achieves state-of-the-art performance on multiple datasets.
arXiv Detail & Related papers (2023-09-25T08:21:49Z) - STRPM: A Spatiotemporal Residual Predictive Model for High-Resolution
Video Prediction [78.129039340528]
We propose a StemporalResidual Predictive Model (STRPM) for high-resolution video prediction.
STRPM can generate more satisfactory results compared with various existing methods.
Experimental results show that STRPM can generate more satisfactory results compared with various existing methods.
arXiv Detail & Related papers (2022-03-30T06:24:00Z) - Memory-Augmented Non-Local Attention for Video Super-Resolution [61.55700315062226]
We propose a novel video super-resolution method that aims at generating high-fidelity high-resolution (HR) videos from low-resolution (LR) ones.
Previous methods predominantly leverage temporal neighbor frames to assist the super-resolution of the current frame.
In contrast, we devise a cross-frame non-local attention mechanism that allows video super-resolution without frame alignment.
arXiv Detail & Related papers (2021-08-25T05:12:14Z) - Video Super-Resolution with Long-Term Self-Exemplars [38.81851816697984]
We propose a video super-resolution method with long-term cross-scale aggregation.
Our model also consists of a multi-reference alignment module to fuse the features derived from similar patches.
To evaluate our proposed method, we conduct extensive experiments on our collected CarCam dataset and the Open dataset.
arXiv Detail & Related papers (2021-06-24T06:07:13Z) - An Efficient Recurrent Adversarial Framework for Unsupervised Real-Time
Video Enhancement [132.60976158877608]
We propose an efficient adversarial video enhancement framework that learns directly from unpaired video examples.
In particular, our framework introduces new recurrent cells that consist of interleaved local and global modules for implicit integration of spatial and temporal information.
The proposed design allows our recurrent cells to efficiently propagate-temporal-information across frames and reduces the need for high complexity networks.
arXiv Detail & Related papers (2020-12-24T00:03:29Z) - Video Super-Resolution with Recurrent Structure-Detail Network [120.1149614834813]
Most video super-resolution methods super-resolve a single reference frame with the help of neighboring frames in a temporal sliding window.
We propose a novel recurrent video super-resolution method which is both effective and efficient in exploiting previous frames to super-resolve the current frame.
arXiv Detail & Related papers (2020-08-02T11:01:19Z) - Transforming Multi-Concept Attention into Video Summarization [36.85535624026879]
We propose a novel attention-based framework for video summarization with complex video data.
Our model can be applied to both labeled and unlabeled data, making our method preferable to real-world applications.
arXiv Detail & Related papers (2020-06-02T06:23:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.