Large Motion Video Super-Resolution with Dual Subnet and Multi-Stage
Communicated Upsampling
- URL: http://arxiv.org/abs/2103.11744v1
- Date: Mon, 22 Mar 2021 11:52:12 GMT
- Title: Large Motion Video Super-Resolution with Dual Subnet and Multi-Stage
Communicated Upsampling
- Authors: Hongying Liu, Peng Zhao, Zhubo Ruan, Fanhua Shang, and Yuanyuan Liu
- Abstract summary: Video super-resolution (VSR) aims at restoring a video in low-resolution (LR) and improving it to higher-resolution (HR)
In this paper, we propose a novel deep neural network with Dual Subnet and Multi-stage Communicated Upsampling (DSMC) for super-resolution of videos with large motion.
- Score: 18.09730129484432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video super-resolution (VSR) aims at restoring a video in low-resolution (LR)
and improving it to higher-resolution (HR). Due to the characteristics of video
tasks, it is very important that motion information among frames should be well
concerned, summarized and utilized for guidance in a VSR algorithm. Especially,
when a video contains large motion, conventional methods easily bring
incoherent results or artifacts. In this paper, we propose a novel deep neural
network with Dual Subnet and Multi-stage Communicated Upsampling (DSMC) for
super-resolution of videos with large motion. We design a new module named
U-shaped residual dense network with 3D convolution (U3D-RDN) for fine implicit
motion estimation and motion compensation (MEMC) as well as coarse spatial
feature extraction. And we present a new Multi-Stage Communicated Upsampling
(MSCU) module to make full use of the intermediate results of upsampling for
guiding the VSR. Moreover, a novel dual subnet is devised to aid the training
of our DSMC, whose dual loss helps to reduce the solution space as well as
enhance the generalization ability. Our experimental results confirm that our
method achieves superior performance on videos with large motion compared to
state-of-the-art methods.
Related papers
- Correlation Matching Transformation Transformers for UHD Image Restoration [46.569124456928535]
This paper proposes a general Transformer for Ultra-High-Definition (UHD) image restoration.
UHDformer contains two learning spaces: (a) learning in high-resolution space and (b) learning in low-resolution space.
Experiments show that our UHDformer reduces about ninety-seven percent model sizes compared with most state-of-the-art methods.
arXiv Detail & Related papers (2024-06-02T06:10:48Z) - Scale-Adaptive Feature Aggregation for Efficient Space-Time Video
Super-Resolution [14.135298731079164]
We propose a novel Scale-Adaptive Feature Aggregation (SAFA) network that adaptively selects sub-networks with different processing scales for individual samples.
Our SAFA network outperforms recent state-of-the-art methods such as TMNet and VideoINR by an average improvement of over 0.5dB on PSNR, while requiring less than half the number of parameters and only 1/3 computational costs.
arXiv Detail & Related papers (2023-10-26T10:18:51Z) - Can SAM Boost Video Super-Resolution? [78.29033914169025]
We propose a simple yet effective module -- SAM-guidEd refinEment Module (SEEM)
This light-weight plug-in module is specifically designed to leverage the attention mechanism for the generation of semantic-aware feature.
We apply our SEEM to two representative methods, EDVR and BasicVSR, resulting in consistently improved performance with minimal implementation effort.
arXiv Detail & Related papers (2023-05-11T02:02:53Z) - Efficient Unsupervised Video Object Segmentation Network Based on Motion
Guidance [1.5736899098702974]
This paper proposes a video object segmentation network based on motion guidance.
The model comprises a dual-stream network, motion guidance module, and multi-scale progressive fusion module.
The experimental results prove the superior performance of the proposed method.
arXiv Detail & Related papers (2022-11-10T06:13:23Z) - Towards Interpretable Video Super-Resolution via Alternating
Optimization [115.85296325037565]
We study a practical space-time video super-resolution (STVSR) problem which aims at generating a high-framerate high-resolution sharp video from a low-framerate blurry video.
We propose an interpretable STVSR framework by leveraging both model-based and learning-based methods.
arXiv Detail & Related papers (2022-07-21T21:34:05Z) - SALISA: Saliency-based Input Sampling for Efficient Video Object
Detection [58.22508131162269]
We propose SALISA, a novel non-uniform SALiency-based Input SAmpling technique for video object detection.
We show that SALISA significantly improves the detection of small objects.
arXiv Detail & Related papers (2022-04-05T17:59:51Z) - Full-Duplex Strategy for Video Object Segmentation [141.43983376262815]
Full- Strategy Network (FSNet) is a novel framework for video object segmentation (VOS)
Our FSNet performs the crossmodal feature-passing (i.e., transmission and receiving) simultaneously before fusion decoding stage.
We show that our FSNet outperforms other state-of-the-arts for both the VOS and video salient object detection tasks.
arXiv Detail & Related papers (2021-08-06T14:50:50Z) - Revisiting Temporal Modeling for Video Super-resolution [47.90584361677039]
We study and compare three temporal modeling methods (2D CNN with early fusion, 3D CNN with slow fusion and Recurrent Neural Network) for video super-resolution.
We also propose a novel Recurrent Residual Network (RRN) for efficient video super-resolution, where residual learning is utilized to stabilize the training of RNN.
arXiv Detail & Related papers (2020-08-13T09:09:37Z) - Zooming Slow-Mo: Fast and Accurate One-Stage Space-Time Video
Super-Resolution [95.26202278535543]
A simple solution is to split it into two sub-tasks: video frame (VFI) and video super-resolution (VSR)
temporalsynthesis and spatial super-resolution are intra-related in this task.
We propose a one-stage space-time video super-resolution framework, which directly synthesizes an HR slow-motion video from an LFR, LR video.
arXiv Detail & Related papers (2020-02-26T16:59:48Z) - Video Face Super-Resolution with Motion-Adaptive Feedback Cell [90.73821618795512]
Video super-resolution (VSR) methods have recently achieved a remarkable success due to the development of deep convolutional neural networks (CNN)
In this paper, we propose a Motion-Adaptive Feedback Cell (MAFC), a simple but effective block, which can efficiently capture the motion compensation and feed it back to the network in an adaptive way.
arXiv Detail & Related papers (2020-02-15T13:14:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.