Learning Event-Based Motion Deblurring
- URL: http://arxiv.org/abs/2004.05794v1
- Date: Mon, 13 Apr 2020 07:01:06 GMT
- Title: Learning Event-Based Motion Deblurring
- Authors: Zhe Jiang, Yu Zhang, Dongqing Zou, Jimmy Ren, Jiancheng Lv, Yebin Liu
- Abstract summary: Fast motion can be captured as events at high time rate for event-based cameras.
We show how its optimization can be unfolded with a novel end-to-end deep architecture.
The proposed approach achieves state-of-the-art reconstruction quality, and generalizes better to handling real-world motion blur.
- Score: 39.16921854492941
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recovering sharp video sequence from a motion-blurred image is highly
ill-posed due to the significant loss of motion information in the blurring
process. For event-based cameras, however, fast motion can be captured as
events at high time rate, raising new opportunities to exploring effective
solutions. In this paper, we start from a sequential formulation of event-based
motion deblurring, then show how its optimization can be unfolded with a novel
end-to-end deep architecture. The proposed architecture is a convolutional
recurrent neural network that integrates visual and temporal knowledge of both
global and local scales in principled manner. To further improve the
reconstruction, we propose a differentiable directional event filtering module
to effectively extract rich boundary prior from the stream of events. We
conduct extensive experiments on the synthetic GoPro dataset and a large newly
introduced dataset captured by a DAVIS240C camera. The proposed approach
achieves state-of-the-art reconstruction quality, and generalizes better to
handling real-world motion blur.
Related papers
- Towards Real-world Event-guided Low-light Video Enhancement and Deblurring [39.942568142125126]
Event cameras have emerged as a promising solution for improving image quality in low-light environments.
We introduce an end-to-end framework to effectively handle these tasks.
Our framework incorporates a module to efficiently leverage temporal information from events and frames.
arXiv Detail & Related papers (2024-08-27T09:44:54Z) - E2HQV: High-Quality Video Generation from Event Camera via
Theory-Inspired Model-Aided Deep Learning [53.63364311738552]
Bio-inspired event cameras or dynamic vision sensors are capable of capturing per-pixel brightness changes (called event-streams) in high temporal resolution and high dynamic range.
It calls for events-to-video (E2V) solutions which take event-streams as input and generate high quality video frames for intuitive visualization.
We propose textbfE2HQV, a novel E2V paradigm designed to produce high-quality video frames from events.
arXiv Detail & Related papers (2024-01-16T05:10:50Z) - EventAid: Benchmarking Event-aided Image/Video Enhancement Algorithms
with Real-captured Hybrid Dataset [55.12137324648253]
Event cameras are emerging imaging technology that offers advantages over conventional frame-based imaging sensors in dynamic range and sensing speed.
This paper focuses on five event-aided image and video enhancement tasks.
arXiv Detail & Related papers (2023-12-13T15:42:04Z) - Learning Parallax for Stereo Event-based Motion Deblurring [8.201943408103995]
Existing approaches rely on the perfect pixel-wise alignment between intensity images and events, which is not always fulfilled in the real world.
We propose a novel coarse-to-fine framework, named NETwork of Event-based motion Deblurring with STereo event and intensity cameras (St-EDNet)
We build a new dataset with STereo Event and Intensity Cameras (StEIC), containing real-world events, intensity images, and dense disparity maps.
arXiv Detail & Related papers (2023-09-18T06:51:41Z) - DynaMoN: Motion-Aware Fast and Robust Camera Localization for Dynamic Neural Radiance Fields [71.94156412354054]
We propose Dynamic Motion-Aware Fast and Robust Camera Localization for Dynamic Neural Radiance Fields (DynaMoN)
DynaMoN handles dynamic content for initial camera pose estimation and statics-focused ray sampling for fast and accurate novel-view synthesis.
We extensively evaluate our approach on two real-world dynamic datasets, the TUM RGB-D dataset and the BONN RGB-D Dynamic dataset.
arXiv Detail & Related papers (2023-09-16T08:46:59Z) - Event-based Non-Rigid Reconstruction from Contours [17.049602518532847]
We propose a novel approach for reconstructing such deformations using measurements from event-based cameras.
Under the assumption of a static background, where all events are generated by the motion, our approach estimates the deformation of objects from events generated at the object contour.
It associates events to mesh faces on the contour and maximizes the alignment of the line of sight through the event pixel with the associated face.
arXiv Detail & Related papers (2022-10-12T14:53:11Z) - Globally-Optimal Event Camera Motion Estimation [30.79931004393174]
Event cameras are bio-inspired sensors that perform well in HDR conditions and have high temporal resolution.
Event cameras measure asynchronous pixel-level changes and return them in a highly discretised format.
arXiv Detail & Related papers (2022-03-08T08:24:22Z) - MEFNet: Multi-scale Event Fusion Network for Motion Deblurring [62.60878284671317]
Traditional frame-based cameras inevitably suffer from motion blur due to long exposure times.
As a kind of bio-inspired camera, the event camera records the intensity changes in an asynchronous way with high temporal resolution.
In this paper, we rethink the event-based image deblurring problem and unfold it into an end-to-end two-stage image restoration network.
arXiv Detail & Related papers (2021-11-30T23:18:35Z) - Motion-blurred Video Interpolation and Extrapolation [72.3254384191509]
We present a novel framework for deblurring, interpolating and extrapolating sharp frames from a motion-blurred video in an end-to-end manner.
To ensure temporal coherence across predicted frames and address potential temporal ambiguity, we propose a simple, yet effective flow-based rule.
arXiv Detail & Related papers (2021-03-04T12:18:25Z) - Back to Event Basics: Self-Supervised Learning of Image Reconstruction
for Event Cameras via Photometric Constancy [0.0]
Event cameras are novel vision sensors that sample, in an asynchronous fashion, brightness increments with low latency and high temporal resolution.
We propose a novel, lightweight neural network for optical flow estimation that achieves high speed inference with only a minor drop in performance.
Results across multiple datasets show that the performance of the proposed self-supervised approach is in line with the state-of-the-art.
arXiv Detail & Related papers (2020-09-17T13:30:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.