Temporal Modulation Network for Controllable Space-Time Video
Super-Resolution
- URL: http://arxiv.org/abs/2104.10642v1
- Date: Wed, 21 Apr 2021 17:10:53 GMT
- Title: Temporal Modulation Network for Controllable Space-Time Video
Super-Resolution
- Authors: Gang Xu and Jun Xu and Zhen Li and Liang Wang and Xing Sun and
Ming-Ming Cheng
- Abstract summary: Space-time video super-resolution aims to increase the spatial and temporal resolutions of low-resolution and low-frame-rate videos.
Deformable convolution based methods have achieved promising STVSR performance, but they could only infer the intermediate frame pre-defined in the training stage.
We propose a Temporal Modulation Network (TMNet) to interpolate arbitrary intermediate frame(s) with accurate high-resolution reconstruction.
- Score: 66.06549492893947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Space-time video super-resolution (STVSR) aims to increase the spatial and
temporal resolutions of low-resolution and low-frame-rate videos. Recently,
deformable convolution based methods have achieved promising STVSR performance,
but they could only infer the intermediate frame pre-defined in the training
stage. Besides, these methods undervalued the short-term motion cues among
adjacent frames. In this paper, we propose a Temporal Modulation Network
(TMNet) to interpolate arbitrary intermediate frame(s) with accurate
high-resolution reconstruction. Specifically, we propose a Temporal Modulation
Block (TMB) to modulate deformable convolution kernels for controllable feature
interpolation. To well exploit the temporal information, we propose a
Locally-temporal Feature Comparison (LFC) module, along with the Bi-directional
Deformable ConvLSTM, to extract short-term and long-term motion cues in videos.
Experiments on three benchmark datasets demonstrate that our TMNet outperforms
previous STVSR methods. The code is available at
https://github.com/CS-GangXu/TMNet.
Related papers
- Global Spatial-Temporal Information-based Residual ConvLSTM for Video Space-Time Super-Resolution [29.74501891293423]
We propose a convolutional neural network (CNN) for space-time video super-resolution, namely GIRNet.
To generate highly accurate features, the proposed network integrates a feature-level temporal module with deformable convolutions and a global spatial-temporal information-based residual convolutional long short-term memory (convLSTM)
Experiments on the Vimeo90K dataset show that the proposed method outperforms state-of-the-art techniques in peak signal-to-noise-ratio (by 1.45 dB, 1.14 dB, and 0.02 dB over STARnet, TMNet, and 3DAttGAN,
arXiv Detail & Related papers (2024-07-11T13:01:44Z) - Continuous Space-Time Video Super-Resolution Utilizing Long-Range
Temporal Information [48.20843501171717]
We propose a continuous ST-VSR (CSTVSR) method that can convert the given video to any frame rate and spatial resolution.
We show that the proposed algorithm has good flexibility and achieves better performance on various datasets.
arXiv Detail & Related papers (2023-02-26T08:02:39Z) - Neighbor Correspondence Matching for Flow-based Video Frame Synthesis [90.14161060260012]
We introduce a neighbor correspondence matching (NCM) algorithm for flow-based frame synthesis.
NCM is performed in a current-frame-agnostic fashion to establish multi-scale correspondences in the spatial-temporal neighborhoods of each pixel.
coarse-scale module is designed to leverage neighbor correspondences to capture large motion, while the fine-scale module is more efficient to speed up the estimation process.
arXiv Detail & Related papers (2022-07-14T09:17:00Z) - RSTT: Real-time Spatial Temporal Transformer for Space-Time Video
Super-Resolution [13.089535703790425]
Space-time video super-resolution (STVSR) is the task of interpolating videos with both Low Frame Rate (LFR) and Low Resolution (LR) to produce High-Frame-Rate (HFR) and also High-Resolution (HR) counterparts.
We propose using a spatial-temporal transformer that naturally incorporates the spatial and temporal super resolution modules into a single model.
arXiv Detail & Related papers (2022-03-27T02:16:26Z) - STDAN: Deformable Attention Network for Space-Time Video
Super-Resolution [39.18399652834573]
We propose a deformable attention network called STDAN for STVSR.
First, we devise a long-short term feature (LSTFI) module, which is capable of abundant content from more neighboring input frames.
Second, we put forward a spatial-temporal deformable feature aggregation (STDFA) module, in which spatial and temporal contexts are adaptively captured and aggregated.
arXiv Detail & Related papers (2022-03-14T03:40:35Z) - Optical-Flow-Reuse-Based Bidirectional Recurrent Network for Space-Time
Video Super-Resolution [52.899234731501075]
Space-time video super-resolution (ST-VSR) simultaneously increases the spatial resolution and frame rate for a given video.
Existing methods typically suffer from difficulties in how to efficiently leverage information from a large range of neighboring frames.
We propose a coarse-to-fine bidirectional recurrent neural network instead of using ConvLSTM to leverage knowledge between adjacent frames.
arXiv Detail & Related papers (2021-10-13T15:21:30Z) - Zooming SlowMo: An Efficient One-Stage Framework for Space-Time Video
Super-Resolution [100.11355888909102]
Space-time video super-resolution aims at generating a high-resolution (HR) slow-motion video from a low-resolution (LR) and low frame rate (LFR) video sequence.
We present a one-stage space-time video super-resolution framework, which can directly reconstruct an HR slow-motion video sequence from an input LR and LFR video.
arXiv Detail & Related papers (2021-04-15T17:59:23Z) - Zooming Slow-Mo: Fast and Accurate One-Stage Space-Time Video
Super-Resolution [95.26202278535543]
A simple solution is to split it into two sub-tasks: video frame (VFI) and video super-resolution (VSR)
temporalsynthesis and spatial super-resolution are intra-related in this task.
We propose a one-stage space-time video super-resolution framework, which directly synthesizes an HR slow-motion video from an LFR, LR video.
arXiv Detail & Related papers (2020-02-26T16:59:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.