EventHands: Real-Time Neural 3D Hand Reconstruction from an Event Stream
- URL: http://arxiv.org/abs/2012.06475v2
- Date: Mon, 14 Dec 2020 16:12:52 GMT
- Title: EventHands: Real-Time Neural 3D Hand Reconstruction from an Event Stream
- Authors: Viktor Rudnev and Vladislav Golyanik and Jiayi Wang and Hans-Peter
Seidel and Franziska Mueller and Mohamed Elgharib and Christian Theobalt
- Abstract summary: 3D hand pose estimation from monocular videos is a long-standing and challenging problem.
We address it for the first time using a single event camera, i.e., an asynchronous vision sensor reacting on brightness changes.
Our approach has characteristics previously not demonstrated with a single RGB or depth camera.
- Score: 80.15360180192175
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D hand pose estimation from monocular videos is a long-standing and
challenging problem, which is now seeing a strong upturn. In this work, we
address it for the first time using a single event camera, i.e., an
asynchronous vision sensor reacting on brightness changes. Our EventHands
approach has characteristics previously not demonstrated with a single RGB or
depth camera such as high temporal resolution at low data throughputs and
real-time performance at 1000 Hz. Due to the different data modality of event
cameras compared to classical cameras, existing methods cannot be directly
applied to and re-trained for event streams. We thus design a new neural
approach which accepts a new event stream representation suitable for learning,
which is trained on newly-generated synthetic event streams and can generalise
to real data. Experiments show that EventHands outperforms recent monocular
methods using a colour (or depth) camera in terms of accuracy and its ability
to capture hand motions of unprecedented speed. Our method, the event stream
simulator and the dataset will be made publicly available.
Related papers
- EventEgo3D: 3D Human Motion Capture from Egocentric Event Streams [59.77837807004765]
This paper introduces a new problem, i.e., 3D human motion capture from an egocentric monocular event camera with a fisheye lens.
Event streams have high temporal resolution and provide reliable cues for 3D human motion capture under high-speed human motions and rapidly changing illumination.
Our EE3D demonstrates robustness and superior 3D accuracy compared to existing solutions while supporting real-time 3D pose update rates of 140Hz.
arXiv Detail & Related papers (2024-04-12T17:59:47Z) - 3D Pose Estimation of Two Interacting Hands from a Monocular Event
Camera [59.846927201816776]
This paper introduces the first framework for 3D tracking of two fast-moving and interacting hands from a single monocular event camera.
Our approach tackles the left-right hand ambiguity with a novel semi-supervised feature-wise attention mechanism and integrates an intersection loss to fix hand collisions.
arXiv Detail & Related papers (2023-12-21T18:59:57Z) - EventNeRF: Neural Radiance Fields from a Single Colour Event Camera [81.19234142730326]
This paper proposes the first approach for 3D-consistent, dense and novel view synthesis using just a single colour event stream as input.
At its core is a neural radiance field trained entirely in a self-supervised manner from events while preserving the original resolution of the colour event channels.
We evaluate our method qualitatively and numerically on several challenging synthetic and real scenes and show that it produces significantly denser and more visually appealing renderings.
arXiv Detail & Related papers (2022-06-23T17:59:53Z) - Differentiable Event Stream Simulator for Non-Rigid 3D Tracking [82.56690776283428]
Our differentiable simulator enables non-rigid 3D tracking of deformable objects from event streams.
We show the effectiveness of our approach for various types of non-rigid objects and compare to existing methods for non-rigid 3D tracking.
arXiv Detail & Related papers (2021-04-30T17:58:07Z) - Learning Monocular Dense Depth from Events [53.078665310545745]
Event cameras produce brightness changes in the form of a stream of asynchronous events instead of intensity frames.
Recent learning-based approaches have been applied to event-based data, such as monocular depth prediction.
We propose a recurrent architecture to solve this task and show significant improvement over standard feed-forward methods.
arXiv Detail & Related papers (2020-10-16T12:36:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.