ShapeAug: Occlusion Augmentation for Event Camera Data
- URL: http://arxiv.org/abs/2401.02274v1
- Date: Thu, 4 Jan 2024 13:49:45 GMT
- Title: ShapeAug: Occlusion Augmentation for Event Camera Data
- Authors: Katharina Bendig, Ren\'e Schuster, Didier Stricker
- Abstract summary: We present a novel event data augmentation approach for Dynamic Vision Sensors (DVSs)
We introduce synthetic events for randomly moving objects in a scene.
We test our method on multiple DVS classification datasets, resulting in an improvement of up to 6.5 % in top1-accuracy.
- Score: 13.634866461329224
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, Dynamic Vision Sensors (DVSs) sparked a lot of interest due to
their inherent advantages over conventional RGB cameras. These advantages
include a low latency, a high dynamic range and a low energy consumption.
Nevertheless, the processing of DVS data using Deep Learning (DL) methods
remains a challenge, particularly since the availability of event training data
is still limited. This leads to a need for event data augmentation techniques
in order to improve accuracy as well as to avoid over-fitting on the training
data. Another challenge especially in real world automotive applications is
occlusion, meaning one object is hindering the view onto the object behind it.
In this paper, we present a novel event data augmentation approach, which
addresses this problem by introducing synthetic events for randomly moving
objects in a scene. We test our method on multiple DVS classification datasets,
resulting in an relative improvement of up to 6.5 % in top1-accuracy. Moreover,
we apply our augmentation technique on the real world Gen1 Automotive Event
Dataset for object detection, where we especially improve the detection of
pedestrians by up to 5 %.
Related papers
- A Recurrent YOLOv8-based framework for Event-Based Object Detection [4.866548300593921]
This study introduces ReYOLOv8, an advanced object detection framework that enhances a frame-based detection system withtemporal modeling capabilities.
We implement a low-latency, memory-efficient method for encoding event data to boost the system's performance.
We also developed a novel data augmentation technique tailored to leverage the unique attributes of event data, thus improving detection accuracy.
arXiv Detail & Related papers (2024-08-09T20:00:16Z) - DailyDVS-200: A Comprehensive Benchmark Dataset for Event-Based Action Recognition [51.96660522869841]
DailyDVS-200 is a benchmark dataset tailored for the event-based action recognition community.
It covers 200 action categories across real-world scenarios, recorded by 47 participants, and comprises more than 22,000 event sequences.
DailyDVS-200 is annotated with 14 attributes, ensuring a detailed characterization of the recorded actions.
arXiv Detail & Related papers (2024-07-06T15:25:10Z) - SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - SPADES: A Realistic Spacecraft Pose Estimation Dataset using Event
Sensing [9.583223655096077]
Due to limited access to real target datasets, algorithms are often trained using synthetic data and applied in the real domain.
Event sensing has been explored in the past and shown to reduce the domain gap between simulations and real-world scenarios.
We introduce a novel dataset, SPADES, comprising real event data acquired in a controlled laboratory environment and simulated event data using the same camera intrinsics.
arXiv Detail & Related papers (2023-11-09T12:14:47Z) - Event-Free Moving Object Segmentation from Moving Ego Vehicle [88.33470650615162]
Moving object segmentation (MOS) in dynamic scenes is an important, challenging, but under-explored research topic for autonomous driving.
Most segmentation methods leverage motion cues obtained from optical flow maps.
We propose to exploit event cameras for better video understanding, which provide rich motion cues without relying on optical flow.
arXiv Detail & Related papers (2023-04-28T23:43:10Z) - Real3D-Aug: Point Cloud Augmentation by Placing Real Objects with
Occlusion Handling for 3D Detection and Segmentation [0.0]
We propose a data augmentation method that takes advantage of already annotated data multiple times.
We propose an augmentation framework that reuses real data, automatically finds suitable placements in the scene to be augmented.
The pipeline proves competitive in training top-performing models for 3D object detection and semantic segmentation.
arXiv Detail & Related papers (2022-06-15T16:25:30Z) - RADA: Robust Adversarial Data Augmentation for Camera Localization in
Challenging Weather [21.198320891744366]
We present RADA, a system that learns to generate minimal image perturbations that are still capable of perplexing the network.
We show that our method achieves up to two times higher accuracy than the SOTA localization models when tested on unseen' challenging weather conditions.
arXiv Detail & Related papers (2021-12-05T03:49:11Z) - DA4Event: towards bridging the Sim-to-Real Gap for Event Cameras using
Domain Adaptation [22.804074390795734]
Event cameras capture pixel-level intensity changes in the form of "events"
The novelty of these sensors results in the lack of a large amount of training data capable of unlocking their potential.
We propose a novel architecture, which better exploits the peculiarities of frame-based event representations.
arXiv Detail & Related papers (2021-03-23T18:09:20Z) - Learning Monocular Dense Depth from Events [53.078665310545745]
Event cameras produce brightness changes in the form of a stream of asynchronous events instead of intensity frames.
Recent learning-based approaches have been applied to event-based data, such as monocular depth prediction.
We propose a recurrent architecture to solve this task and show significant improvement over standard feed-forward methods.
arXiv Detail & Related papers (2020-10-16T12:36:23Z) - SCRDet++: Detecting Small, Cluttered and Rotated Objects via
Instance-Level Feature Denoising and Rotation Loss Smoothing [131.04304632759033]
Small and cluttered objects are common in real-world which are challenging for detection.
In this paper, we first innovatively introduce the idea of denoising to object detection.
Instance-level denoising on the feature map is performed to enhance the detection to small and cluttered objects.
arXiv Detail & Related papers (2020-04-28T06:03:54Z) - Quantifying Data Augmentation for LiDAR based 3D Object Detection [139.64869289514525]
In this work, we shed light on different data augmentation techniques commonly used in Light Detection and Ranging (LiDAR) based 3D Object Detection.
We investigate a variety of global and local augmentation techniques, where global augmentation techniques are applied to the entire point cloud of a scene and local augmentation techniques are only applied to points belonging to individual objects in the scene.
Our findings show that both types of data augmentation can lead to performance increases, but it also turns out, that some augmentation techniques, such as individual object translation, for example, can be counterproductive and can hurt the overall performance.
arXiv Detail & Related papers (2020-04-03T16:09:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.