Feature-Aligned Video Raindrop Removal with Temporal Constraints
- URL: http://arxiv.org/abs/2205.14574v1
- Date: Sun, 29 May 2022 05:42:14 GMT
- Title: Feature-Aligned Video Raindrop Removal with Temporal Constraints
- Authors: Wending Yan, Lu Xu, Wenhan Yang and Robby T. Tan
- Abstract summary: Raindrop removal is challenging for both single image and video.
Unlike rain streaks, adherent raindrops tend to cover the same area in several frames.
Our method employs a two-stage video-based raindrop removal method.
- Score: 68.49161092870224
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing adherent raindrop removal methods focus on the detection of the
raindrop locations, and then use inpainting techniques or generative networks
to recover the background behind raindrops. Yet, as adherent raindrops are
diverse in sizes and appearances, the detection is challenging for both single
image and video. Moreover, unlike rain streaks, adherent raindrops tend to
cover the same area in several frames. Addressing these problems, our method
employs a two-stage video-based raindrop removal method. The first stage is the
single image module, which generates initial clean results. The second stage is
the multiple frame module, which further refines the initial results using
temporal constraints, namely, by utilizing multiple input frames in our process
and applying temporal consistency between adjacent output frames. Our single
image module employs a raindrop removal network to generate initial raindrop
removal results, and create a mask representing the differences between the
input and initial output. Once the masks and initial results for consecutive
frames are obtained, our multiple-frame module aligns the frames in both the
image and feature levels and then obtains the clean background. Our method
initially employs optical flow to align the frames, and then utilizes
deformable convolution layers further to achieve feature-level frame alignment.
To remove small raindrops and recover correct backgrounds, a target frame is
predicted from adjacent frames. A series of unsupervised losses are proposed so
that our second stage, which is the video raindrop removal module, can
self-learn from video data without ground truths. Experimental results on real
videos demonstrate the state-of-art performance of our method both
quantitatively and qualitatively.
Related papers
- RainyScape: Unsupervised Rainy Scene Reconstruction using Decoupled Neural Rendering [50.14860376758962]
We propose RainyScape, an unsupervised framework for reconstructing clean scenes from a collection of multi-view rainy images.
Based on the spectral bias property of neural networks, we first optimize the neural rendering pipeline to obtain a low-frequency scene representation.
We jointly optimize the two modules, driven by the proposed adaptive direction-sensitive gradient-based reconstruction loss.
arXiv Detail & Related papers (2024-04-17T14:07:22Z) - EGVD: Event-Guided Video Deraining [57.59935209162314]
We propose an end-to-end learning-based network to unlock the potential of the event camera for video deraining.
We build a real-world dataset consisting of rainy videos and temporally synchronized event streams.
arXiv Detail & Related papers (2023-09-29T13:47:53Z) - Beyond Monocular Deraining: Parallel Stereo Deraining Network Via
Semantic Prior [103.49307603952144]
Most existing de-rain algorithms use only one single input image and aim to recover a clean image.
We present a Paired Rain Removal Network (PRRNet), which exploits both stereo images and semantic information.
Experiments on both monocular and the newly proposed stereo rainy datasets demonstrate that the proposed method achieves the state-of-the-art performance.
arXiv Detail & Related papers (2021-05-09T04:15:10Z) - Restoration of Video Frames from a Single Blurred Image with Motion
Understanding [69.90724075337194]
We propose a novel framework to generate clean video frames from a single motion-red image.
We formulate video restoration from a single blurred image as an inverse problem by setting clean image sequence and their respective motion as latent factors.
Our framework is based on anblur-decoder structure with spatial transformer network modules.
arXiv Detail & Related papers (2021-04-19T08:32:57Z) - Motion-blurred Video Interpolation and Extrapolation [72.3254384191509]
We present a novel framework for deblurring, interpolating and extrapolating sharp frames from a motion-blurred video in an end-to-end manner.
To ensure temporal coherence across predicted frames and address potential temporal ambiguity, we propose a simple, yet effective flow-based rule.
arXiv Detail & Related papers (2021-03-04T12:18:25Z) - Multi-Stage Raw Video Denoising with Adversarial Loss and Gradient Mask [14.265454188161819]
We propose a learning-based approach for denoising raw videos captured under low lighting conditions.
We first explicitly align the neighboring frames to the current frame using a convolutional neural network (CNN)
We then fuse the registered frames using another CNN to obtain the final denoised frame.
arXiv Detail & Related papers (2021-03-04T06:57:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.