DA4Event: towards bridging the Sim-to-Real Gap for Event Cameras using
Domain Adaptation
- URL: http://arxiv.org/abs/2103.12768v1
- Date: Tue, 23 Mar 2021 18:09:20 GMT
- Title: DA4Event: towards bridging the Sim-to-Real Gap for Event Cameras using
Domain Adaptation
- Authors: Mirco Planamente and Chiara Plizzari and Marco Cannici and Marco
Ciccone and Francesco Strada and Andrea Bottino and Matteo Matteucci and
Barbara Caputo
- Abstract summary: Event cameras capture pixel-level intensity changes in the form of "events"
The novelty of these sensors results in the lack of a large amount of training data capable of unlocking their potential.
We propose a novel architecture, which better exploits the peculiarities of frame-based event representations.
- Score: 22.804074390795734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Event cameras are novel bio-inspired sensors, which asynchronously capture
pixel-level intensity changes in the form of "events". The innovative way they
acquire data presents several advantages over standard devices, especially in
poor lighting and high-speed motion conditions. However, the novelty of these
sensors results in the lack of a large amount of training data capable of fully
unlocking their potential. The most common approach implemented by researchers
to address this issue is to leverage simulated event data. Yet, this approach
comes with an open research question: how well simulated data generalize to
real data? To answer this, we propose to exploit, in the event-based context,
recent Domain Adaptation (DA) advances in traditional computer vision, showing
that DA techniques applied to event data help reduce the sim-to-real gap. To
this purpose, we propose a novel architecture, which we call Multi-View DA4E
(MV-DA4E), that better exploits the peculiarities of frame-based event
representations while also promoting domain invariant characteristics in
features. Through extensive experiments, we prove the effectiveness of DA
methods and MV-DA4E on N-Caltech101. Moreover, we validate their soundness in a
real-world scenario through a cross-domain analysis on the popular RGB-D Object
Dataset (ROD), which we extended to the event modality (RGB-E).
Related papers
- How Important are Data Augmentations to Close the Domain Gap for Object Detection in Orbit? [15.550663626482903]
We investigate the efficacy of data augmentations to close the domain gap in spaceborne computer vision.
We propose two novel data augmentations specifically developed to emulate the visual effects observed in orbital imagery.
arXiv Detail & Related papers (2024-10-21T08:24:46Z) - DailyDVS-200: A Comprehensive Benchmark Dataset for Event-Based Action Recognition [51.96660522869841]
DailyDVS-200 is a benchmark dataset tailored for the event-based action recognition community.
It covers 200 action categories across real-world scenarios, recorded by 47 participants, and comprises more than 22,000 event sequences.
DailyDVS-200 is annotated with 14 attributes, ensuring a detailed characterization of the recorded actions.
arXiv Detail & Related papers (2024-07-06T15:25:10Z) - RaSim: A Range-aware High-fidelity RGB-D Data Simulation Pipeline for Real-world Applications [55.24463002889]
We focus on depth data synthesis and develop a range-aware RGB-D data simulation pipeline (RaSim)
In particular, high-fidelity depth data is generated by imitating the imaging principle of real-world sensors.
RaSim can be directly applied to real-world scenarios without any finetuning and excel at downstream RGB-D perception tasks.
arXiv Detail & Related papers (2024-04-05T08:52:32Z) - ShapeAug: Occlusion Augmentation for Event Camera Data [13.634866461329224]
We present a novel event data augmentation approach for Dynamic Vision Sensors (DVSs)
We introduce synthetic events for randomly moving objects in a scene.
We test our method on multiple DVS classification datasets, resulting in an improvement of up to 6.5 % in top1-accuracy.
arXiv Detail & Related papers (2024-01-04T13:49:45Z) - Implicit Event-RGBD Neural SLAM [54.74363487009845]
Implicit neural SLAM has achieved remarkable progress recently.
Existing methods face significant challenges in non-ideal scenarios.
We propose EN-SLAM, the first event-RGBD implicit neural SLAM framework.
arXiv Detail & Related papers (2023-11-18T08:48:58Z) - SPADES: A Realistic Spacecraft Pose Estimation Dataset using Event
Sensing [9.583223655096077]
Due to limited access to real target datasets, algorithms are often trained using synthetic data and applied in the real domain.
Event sensing has been explored in the past and shown to reduce the domain gap between simulations and real-world scenarios.
We introduce a novel dataset, SPADES, comprising real event data acquired in a controlled laboratory environment and simulated event data using the same camera intrinsics.
arXiv Detail & Related papers (2023-11-09T12:14:47Z) - Deformable Neural Radiance Fields using RGB and Event Cameras [65.40527279809474]
We develop a novel method to model the deformable neural radiance fields using RGB and event cameras.
The proposed method uses the asynchronous stream of events and sparse RGB frames.
Experiments conducted on both realistically rendered graphics and real-world datasets demonstrate a significant benefit of the proposed method.
arXiv Detail & Related papers (2023-09-15T14:19:36Z) - EventTransAct: A video transformer-based framework for Event-camera
based action recognition [52.537021302246664]
Event cameras offer new opportunities compared to standard action recognition in RGB videos.
In this study, we employ a computationally efficient model, namely the video transformer network (VTN), which initially acquires spatial embeddings per event-frame.
In order to better adopt the VTN for the sparse and fine-grained nature of event data, we design Event-Contrastive Loss ($mathcalL_EC$) and event-specific augmentations.
arXiv Detail & Related papers (2023-08-25T23:51:07Z) - Learning Monocular Dense Depth from Events [53.078665310545745]
Event cameras produce brightness changes in the form of a stream of asynchronous events instead of intensity frames.
Recent learning-based approaches have been applied to event-based data, such as monocular depth prediction.
We propose a recurrent architecture to solve this task and show significant improvement over standard feed-forward methods.
arXiv Detail & Related papers (2020-10-16T12:36:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.