Revitalizing Legacy Video Content: Deinterlacing with Bidirectional
Information Propagation
- URL: http://arxiv.org/abs/2310.19535v2
- Date: Tue, 5 Dec 2023 15:06:02 GMT
- Title: Revitalizing Legacy Video Content: Deinterlacing with Bidirectional
Information Propagation
- Authors: Zhaowei Gao, Mingyang Song, Christopher Schroers, Yang Zhang
- Abstract summary: We present a deep-learning-based method for deinterlacing animated and live-action video content.
Our proposed method supports bidirectional-temporal information propagation across multiple scales.
Our method can process multiple fields simultaneously, reducing per-frame time, and potentially enabling real-time processing.
- Score: 14.340811078427553
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to old CRT display technology and limited transmission bandwidth, early
film and TV broadcasts commonly used interlaced scanning. This meant each field
contained only half of the information. Since modern displays require full
frames, this has spurred research into deinterlacing, i.e. restoring the
missing information in legacy video content. In this paper, we present a
deep-learning-based method for deinterlacing animated and live-action content.
Our proposed method supports bidirectional spatio-temporal information
propagation across multiple scales to leverage information in both space and
time. More specifically, we design a Flow-guided Refinement Block (FRB) which
performs feature refinement including alignment, fusion, and rectification.
Additionally, our method can process multiple fields simultaneously, reducing
per-frame processing time, and potentially enabling real-time processing. Our
experimental results demonstrate that our proposed method achieves superior
performance compared to existing methods.
Related papers
- CMTA: Cross-Modal Temporal Alignment for Event-guided Video Deblurring [44.30048301161034]
Video deblurring aims to enhance the quality of restored results in motion-red videos by gathering information from adjacent video frames.
We propose two modules: 1) Intra-frame feature enhancement operates within the exposure time of a single blurred frame, and 2) Inter-frame temporal feature alignment gathers valuable long-range temporal information to target frames.
We demonstrate that our proposed methods outperform state-of-the-art frame-based and event-based motion deblurring methods through extensive experiments conducted on both synthetic and real-world deblurring datasets.
arXiv Detail & Related papers (2024-08-27T10:09:17Z) - COVE: Unleashing the Diffusion Feature Correspondence for Consistent Video Editing [57.76170824395532]
Video editing is an emerging task, in which most current methods adopt the pre-trained text-to-image (T2I) diffusion model to edit the source video.
We propose COrrespondence-guided Video Editing (COVE) to achieve high-quality and consistent video editing.
COVE can be seamlessly integrated into the pre-trained T2I diffusion model without the need for extra training or optimization.
arXiv Detail & Related papers (2024-06-13T06:27:13Z) - Blur-aware Spatio-temporal Sparse Transformer for Video Deblurring [14.839956958725883]
We propose blurbfBSSTNet, textbfBlur-aware textbfStext-temporal textbfTransformer Network.
The proposed BSSTNet outperforms the state-of-the-art methods on the GoPro and DVD datasets.
arXiv Detail & Related papers (2024-06-11T17:59:56Z) - Collaborative Feedback Discriminative Propagation for Video Super-Resolution [66.61201445650323]
Key success of video super-resolution (VSR) methods stems mainly from exploring spatial and temporal information.
Inaccurate alignment usually leads to aligned features with significant artifacts.
propagation modules only propagate the same timestep features forward or backward.
arXiv Detail & Related papers (2024-04-06T22:08:20Z) - Continuous Space-Time Video Super-Resolution Utilizing Long-Range
Temporal Information [48.20843501171717]
We propose a continuous ST-VSR (CSTVSR) method that can convert the given video to any frame rate and spatial resolution.
We show that the proposed algorithm has good flexibility and achieves better performance on various datasets.
arXiv Detail & Related papers (2023-02-26T08:02:39Z) - Deep Video Prior for Video Consistency and Propagation [58.250209011891904]
We present a novel and general approach for blind video temporal consistency.
Our method is only trained on a pair of original and processed videos directly instead of a large dataset.
We show that temporal consistency can be achieved by training a convolutional neural network on a video with Deep Video Prior.
arXiv Detail & Related papers (2022-01-27T16:38:52Z) - Deep Recurrent Neural Network with Multi-scale Bi-directional
Propagation for Video Deblurring [36.94523101375519]
We propose a deep Recurrent Neural Network with Multi-scale Bi-directional Propagation (RNN-MBP) to propagate and gather information from unaligned neighboring frames for better video deblurring.
To better evaluate the proposed algorithm and existing state-of-the-art methods on real-world blurry scenes, we also create a Real-World Blurry Video dataset.
The proposed algorithm performs favorably against the state-of-the-art methods on three typical benchmarks.
arXiv Detail & Related papers (2021-12-09T11:02:56Z) - Optical-Flow-Reuse-Based Bidirectional Recurrent Network for Space-Time
Video Super-Resolution [52.899234731501075]
Space-time video super-resolution (ST-VSR) simultaneously increases the spatial resolution and frame rate for a given video.
Existing methods typically suffer from difficulties in how to efficiently leverage information from a large range of neighboring frames.
We propose a coarse-to-fine bidirectional recurrent neural network instead of using ConvLSTM to leverage knowledge between adjacent frames.
arXiv Detail & Related papers (2021-10-13T15:21:30Z) - Spatiotemporal Inconsistency Learning for DeepFake Video Detection [51.747219106855624]
We present a novel temporal modeling paradigm in TIM by exploiting the temporal difference over adjacent frames along with both horizontal and vertical directions.
And the ISM simultaneously utilizes the spatial information from SIM and temporal information from TIM to establish a more comprehensive spatial-temporal representation.
arXiv Detail & Related papers (2021-09-04T13:05:37Z) - Coarse-Fine Networks for Temporal Activity Detection in Videos [45.03545172714305]
We introduce 'Co-Fine Networks', a two-stream architecture which benefits from different abstractions of temporal resolution to learn better video representations for long-term motion.
We show that our method can outperform the state-of-the-arts for action detection in public datasets with a significantly reduced compute and memory footprint.
arXiv Detail & Related papers (2021-03-01T20:48:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.