Progressive Deep Video Dehazing without Explicit Alignment Estimation
- URL: http://arxiv.org/abs/2107.07837v1
- Date: Fri, 16 Jul 2021 11:57:40 GMT
- Title: Progressive Deep Video Dehazing without Explicit Alignment Estimation
- Authors: Runde Li
- Abstract summary: We propose a progressive alignment and restoration method for video dehazing.
The alignment process aligns consecutive neighboring frames stage by stage without using the optical flow estimation.
The restoration process is not only implemented under the alignment process but also uses a refinement network to improve the dehazing performance of the whole network.
- Score: 2.766648389933265
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To solve the issue of video dehazing, there are two main tasks to attain: how
to align adjacent frames to the reference frame; how to restore the reference
frame. Some papers adopt explicit approaches (e.g., the Markov random field,
optical flow, deformable convolution, 3D convolution) to align neighboring
frames with the reference frame in feature space or image space, they then use
various restoration methods to achieve the final dehazing results. In this
paper, we propose a progressive alignment and restoration method for video
dehazing. The alignment process aligns consecutive neighboring frames stage by
stage without using the optical flow estimation. The restoration process is not
only implemented under the alignment process but also uses a refinement network
to improve the dehazing performance of the whole network. The proposed networks
include four fusion networks and one refinement network. To decrease the
parameters of networks, three fusion networks in the first fusion stage share
the same parameters. Extensive experiments demonstrate that the proposed video
dehazing method achieves outstanding performance against the-state-of-art
methods.
Related papers
- Video Deblurring with Deconvolution and Aggregation Networks [1.6114012813668932]
We propose a deconvolution and aggregation network (DAN) for video deblurring.<n>In DAN, both deconvolution and aggregation strategies are achieved through three sub-networks.<n>The proper combination of three sub-networks can achieve favorable performance on video deblurring by using the neighbor frames suitably.
arXiv Detail & Related papers (2025-06-04T15:19:11Z) - CoordFlow: Coordinate Flow for Pixel-wise Neural Video Representation [11.364753833652182]
Implicit Neural Representation (INR) is a promising alternative to traditional transform-based methodologies.
We introduce CoordFlow, a novel pixel-wise INR for video compression.
It yields state-of-the-art results compared to other pixel-wise INRs and on-par performance compared to leading frame-wise techniques.
arXiv Detail & Related papers (2025-01-01T22:58:06Z) - Optical-Flow Guided Prompt Optimization for Coherent Video Generation [51.430833518070145]
We propose a framework called MotionPrompt that guides the video generation process via optical flow.
We optimize learnable token embeddings during reverse sampling steps by using gradients from a trained discriminator applied to random frame pairs.
This approach allows our method to generate visually coherent video sequences that closely reflect natural motion dynamics, without compromising the fidelity of the generated content.
arXiv Detail & Related papers (2024-11-23T12:26:52Z) - Motion-Aware Video Frame Interpolation [49.49668436390514]
We introduce a Motion-Aware Video Frame Interpolation (MA-VFI) network, which directly estimates intermediate optical flow from consecutive frames.
It not only extracts global semantic relationships and spatial details from input frames with different receptive fields, but also effectively reduces the required computational cost and complexity.
arXiv Detail & Related papers (2024-02-05T11:00:14Z) - Training-Free Semantic Video Composition via Pre-trained Diffusion Model [96.0168609879295]
Current approaches, predominantly trained on videos with adjusted foreground color and lighting, struggle to address deep semantic disparities beyond superficial adjustments.
We propose a training-free pipeline employing a pre-trained diffusion model imbued with semantic prior knowledge.
Experimental results reveal that our pipeline successfully ensures the visual harmony and inter-frame coherence of the outputs.
arXiv Detail & Related papers (2024-01-17T13:07:22Z) - Aggregating Nearest Sharp Features via Hybrid Transformers for Video Deblurring [70.06559269075352]
We propose a video deblurring method that leverages both neighboring frames and existing sharp frames using hybrid Transformers for feature aggregation.
To aggregate nearest sharp features from detected sharp frames, we utilize a global Transformer with multi-scale matching capability.
Our proposed method outperforms state-of-the-art video deblurring methods as well as event-driven video deblurring methods in terms of quantitative metrics and visual quality.
arXiv Detail & Related papers (2023-09-13T16:12:11Z) - Meta-Auxiliary Network for 3D GAN Inversion [18.777352198191004]
In this work, we present a novel meta-auxiliary framework, while leveraging the newly developed 3D GANs as generator.
In the first stage, we invert the input image to an editable latent code using off-the-shelf inversion techniques.
The auxiliary network is proposed to refine the generator parameters with the given image as input, which both predicts offsets for weights of convolutional layers and sampling positions of volume rendering.
In the second stage, we perform meta-learning to fast adapt the auxiliary network to the input image, then the final reconstructed image is synthesized via the meta-learned auxiliary network.
arXiv Detail & Related papers (2023-05-18T11:26:27Z) - TimeLens: Event-based Video Frame Interpolation [54.28139783383213]
We introduce Time Lens, a novel indicates equal contribution method that leverages the advantages of both synthesis-based and flow-based approaches.
We show an up to 5.21 dB improvement in terms of PSNR over state-of-the-art frame-based and event-based methods.
arXiv Detail & Related papers (2021-06-14T10:33:47Z) - EA-Net: Edge-Aware Network for Flow-based Video Frame Interpolation [101.75999290175412]
We propose to reduce the image blur and get the clear shape of objects by preserving the edges in the interpolated frames.
The proposed Edge-Aware Network (EANet) integrates the edge information into the frame task.
Three edge-aware mechanisms are developed to emphasize the frame edges in estimating flow maps.
arXiv Detail & Related papers (2021-05-17T08:44:34Z) - FDAN: Flow-guided Deformable Alignment Network for Video
Super-Resolution [12.844337773258678]
Flow-guided Deformable Module (FDM) is proposed to integrate optical flow into deformable convolution.
FDAN reaches the state-of-the-art performance on two benchmark datasets.
arXiv Detail & Related papers (2021-05-12T13:18:36Z) - Video Frame Interpolation via Structure-Motion based Iterative Fusion [19.499969588931414]
We propose a structure-motion based iterative fusion method for video frame Interpolation.
Inspired by the observation that audiences have different visual preferences on foreground and background objects, we for the first time propose to use saliency masks in the evaluation processes of the task of video frame Interpolation.
arXiv Detail & Related papers (2021-05-11T22:11:17Z) - Restoration of Video Frames from a Single Blurred Image with Motion
Understanding [69.90724075337194]
We propose a novel framework to generate clean video frames from a single motion-red image.
We formulate video restoration from a single blurred image as an inverse problem by setting clean image sequence and their respective motion as latent factors.
Our framework is based on anblur-decoder structure with spatial transformer network modules.
arXiv Detail & Related papers (2021-04-19T08:32:57Z) - Multi-Stage Raw Video Denoising with Adversarial Loss and Gradient Mask [14.265454188161819]
We propose a learning-based approach for denoising raw videos captured under low lighting conditions.
We first explicitly align the neighboring frames to the current frame using a convolutional neural network (CNN)
We then fuse the registered frames using another CNN to obtain the final denoised frame.
arXiv Detail & Related papers (2021-03-04T06:57:48Z) - FineNet: Frame Interpolation and Enhancement for Face Video Deblurring [18.49184807837449]
The aim of this work is to deblur face videos.
We propose a method that tackles this problem from two directions: (1) enhancing the blurry frames, and (2) treating the blurry frames as missing values and estimate them by objective.
Experiments on three real and synthetically generated video datasets show that our method outperforms the previous state-of-the-art methods by a large margin in terms of both quantitative and qualitative results.
arXiv Detail & Related papers (2021-03-01T09:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.