Across Scales & Across Dimensions: Temporal Super-Resolution using Deep
Internal Learning
- URL: http://arxiv.org/abs/2003.08872v3
- Date: Thu, 15 Oct 2020 10:09:29 GMT
- Title: Across Scales & Across Dimensions: Temporal Super-Resolution using Deep
Internal Learning
- Authors: Liad Pollak Zuckerman, Eyal Naor, George Pisha, Shai Bagon, Michal
Irani
- Abstract summary: We train a video-specific CNN on examples extracted directly from the low-framerate input video.
Our method exploits the strong recurrence of small space-time patches inside a single video sequence.
The higher spatial resolution of video frames provides strong examples as to how to increase the temporal temporal resolution of that video.
- Score: 11.658606722158517
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: When a very fast dynamic event is recorded with a low-framerate camera, the
resulting video suffers from severe motion blur (due to exposure time) and
motion aliasing (due to low sampling rate in time). True Temporal
Super-Resolution (TSR) is more than just Temporal-Interpolation (increasing
framerate). It can also recover new high temporal frequencies beyond the
temporal Nyquist limit of the input video, thus resolving both motion-blur and
motion-aliasing effects that temporal frame interpolation (as sophisticated as
it maybe) cannot undo. In this paper we propose a "Deep Internal Learning"
approach for true TSR. We train a video-specific CNN on examples extracted
directly from the low-framerate input video. Our method exploits the strong
recurrence of small space-time patches inside a single video sequence, both
within and across different spatio-temporal scales of the video. We further
observe (for the first time) that small space-time patches recur also
across-dimensions of the video sequence - i.e., by swapping the spatial and
temporal dimensions. In particular, the higher spatial resolution of video
frames provides strong examples as to how to increase the temporal resolution
of that video. Such internal video-specific examples give rise to strong
self-supervision, requiring no data but the input video itself. This results in
Zero-Shot Temporal-SR of complex videos, which removes both motion blur and
motion aliasing, outperforming previous supervised methods trained on external
video datasets.
Related papers
- Towards Interpretable Video Super-Resolution via Alternating
Optimization [115.85296325037565]
We study a practical space-time video super-resolution (STVSR) problem which aims at generating a high-framerate high-resolution sharp video from a low-framerate blurry video.
We propose an interpretable STVSR framework by leveraging both model-based and learning-based methods.
arXiv Detail & Related papers (2022-07-21T21:34:05Z) - VideoINR: Learning Video Implicit Neural Representation for Continuous
Space-Time Super-Resolution [75.79379734567604]
We show that Video Implicit Neural Representation (VideoINR) can be decoded to videos of arbitrary spatial resolution and frame rate.
We show that VideoINR achieves competitive performances with state-of-the-art STVSR methods on common up-sampling scales.
arXiv Detail & Related papers (2022-06-09T17:45:49Z) - Unsupervised Video Interpolation by Learning Multilayered 2.5D Motion
Fields [75.81417944207806]
This paper presents a self-supervised approach to video frame learning that requires only a single video.
We parameterize the video motions by solving an ordinary differentiable equation (ODE) defined on a time-varying motion field.
This implicit neural representation learns the video as a space-time continuum, allowing frame-time continuum at any temporal resolution.
arXiv Detail & Related papers (2022-04-21T06:17:05Z) - Learning Spatio-Temporal Downsampling for Effective Video Upscaling [20.07194339353278]
In this paper, we aim to solve the space-time aliasing problem by learning a-temporal downsampling and upsampling.
Our framework enables a variety of applications, including arbitrary video resampling, blurry frame reconstruction, and efficient video storage.
arXiv Detail & Related papers (2022-03-15T17:59:00Z) - Video Frame Interpolation without Temporal Priors [91.04877640089053]
Video frame aims to synthesize non-exist intermediate frames in a video sequence.
The temporal priors of videos, i.e. frames per second (FPS) and frame exposure time, may vary from different camera sensors.
We devise a novel optical flow refinement strategy for better synthesizing results.
arXiv Detail & Related papers (2021-12-02T12:13:56Z) - Zooming SlowMo: An Efficient One-Stage Framework for Space-Time Video
Super-Resolution [100.11355888909102]
Space-time video super-resolution aims at generating a high-resolution (HR) slow-motion video from a low-resolution (LR) and low frame rate (LFR) video sequence.
We present a one-stage space-time video super-resolution framework, which can directly reconstruct an HR slow-motion video sequence from an input LR and LFR video.
arXiv Detail & Related papers (2021-04-15T17:59:23Z) - Zooming Slow-Mo: Fast and Accurate One-Stage Space-Time Video
Super-Resolution [95.26202278535543]
A simple solution is to split it into two sub-tasks: video frame (VFI) and video super-resolution (VSR)
temporalsynthesis and spatial super-resolution are intra-related in this task.
We propose a one-stage space-time video super-resolution framework, which directly synthesizes an HR slow-motion video from an LFR, LR video.
arXiv Detail & Related papers (2020-02-26T16:59:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.