Artificial Dummies for Urban Dataset Augmentation
- URL: http://arxiv.org/abs/2012.08274v1
- Date: Tue, 15 Dec 2020 13:17:25 GMT
- Title: Artificial Dummies for Urban Dataset Augmentation
- Authors: Anton\'in Vobeck\'y, David Hurych, Michal U\v{r}i\v{c}\'a\v{r},
Patrick P\'erez, and Josef \v{S}ivic
- Abstract summary: Existing datasets for training pedestrian detectors in images suffer from limited appearance and pose variation.
This paper describes an augmentation method for controlled synthesis of urban scenes containing people.
We demonstrate that the data generated by our DummyNet improve performance of several existing person detectors across various datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing datasets for training pedestrian detectors in images suffer from
limited appearance and pose variation. The most challenging scenarios are
rarely included because they are too difficult to capture due to safety
reasons, or they are very unlikely to happen. The strict safety requirements in
assisted and autonomous driving applications call for an extra high detection
accuracy also in these rare situations. Having the ability to generate people
images in arbitrary poses, with arbitrary appearances and embedded in different
background scenes with varying illumination and weather conditions, is a
crucial component for the development and testing of such applications. The
contributions of this paper are three-fold. First, we describe an augmentation
method for controlled synthesis of urban scenes containing people, thus
producing rare or never-seen situations. This is achieved with a data generator
(called DummyNet) with disentangled control of the pose, the appearance, and
the target background scene. Second, the proposed generator relies on novel
network architecture and associated loss that takes into account the
segmentation of the foreground person and its composition into the background
scene. Finally, we demonstrate that the data generated by our DummyNet improve
performance of several existing person detectors across various datasets as
well as in challenging situations, such as night-time conditions, where only a
limited amount of training data is available. In the setup with only day-time
data available, we improve the night-time detector by $17\%$ log-average miss
rate over the detector trained with the day-time data only.
Related papers
- Multi-Attention Fusion Drowsy Driving Detection Model [1.2043574473965317]
We introduce a novel approach called the Multi-Attention Fusion Drowsy Driving Detection Model (MAF)
Our proposed model achieves an impressive driver drowsiness detection accuracy of 96.8%.
arXiv Detail & Related papers (2023-12-28T14:53:32Z) - MonoTDP: Twin Depth Perception for Monocular 3D Object Detection in
Adverse Scenes [49.21187418886508]
This paper proposes a monocular 3D detection model designed to perceive twin depth in adverse scenes, termed MonoTDP.
We first introduce an adaptive learning strategy to aid the model in handling uncontrollable weather conditions, significantly resisting degradation caused by various degrading factors.
Then, to address the depth/content loss in adverse regions, we propose a novel twin depth perception module that simultaneously estimates scene and object depth.
arXiv Detail & Related papers (2023-05-18T13:42:02Z) - ScatterNeRF: Seeing Through Fog with Physically-Based Inverse Neural
Rendering [83.75284107397003]
We introduce ScatterNeRF, a neural rendering method which renders scenes and decomposes the fog-free background.
We propose a disentangled representation for the scattering volume and the scene objects, and learn the scene reconstruction with physics-inspired losses.
We validate our method by capturing multi-view In-the-Wild data and controlled captures in a large-scale fog chamber.
arXiv Detail & Related papers (2023-05-03T13:24:06Z) - Generalized Few-Shot 3D Object Detection of LiDAR Point Cloud for
Autonomous Driving [91.39625612027386]
We propose a novel task, called generalized few-shot 3D object detection, where we have a large amount of training data for common (base) objects, but only a few data for rare (novel) classes.
Specifically, we analyze in-depth differences between images and point clouds, and then present a practical principle for the few-shot setting in the 3D LiDAR dataset.
To solve this task, we propose an incremental fine-tuning method to extend existing 3D detection models to recognize both common and rare objects.
arXiv Detail & Related papers (2023-02-08T07:11:36Z) - Adapting to Skew: Imputing Spatiotemporal Urban Data with 3D Partial
Convolutions and Biased Masking [13.94102520443797]
Missing regions in urban data can be caused by sensor or software failures, data quality issues, interference from weather events, incomplete data collection, or varying data use regulations.
We adapt computer vision techniques for image inpainting to operate on 3D histograms (2D space + 1D time) commonly used for data exchange in urban settings.
We show that the core model is effective qualitatively and quantitatively, and that biased masking during training reduces error in a variety of scenarios.
arXiv Detail & Related papers (2023-01-10T22:44:22Z) - Ithaca365: Dataset and Driving Perception under Repeated and Challenging
Weather Conditions [0.0]
We present a new dataset to enable robust autonomous driving via a novel data collection process.
The dataset includes images and point clouds from cameras and LiDAR sensors, along with high-precision GPS/INS.
We demonstrate the uniqueness of this dataset by analyzing the performance of baselines in amodal segmentation of road and objects.
arXiv Detail & Related papers (2022-08-01T22:55:32Z) - Neural-Sim: Learning to Generate Training Data with NeRF [31.81496344354997]
We present the first fully differentiable synthetic data pipeline that uses Neural Radiance Fields (NeRFs) in a closed-loop with a target application's loss function.
Our approach generates data on-demand, with no human labor, to maximize accuracy for a target task.
arXiv Detail & Related papers (2022-07-22T22:48:33Z) - CrowdDriven: A New Challenging Dataset for Outdoor Visual Localization [44.97567243883994]
We propose a new benchmark for visual localization in outdoor scenes using crowd-sourced data.
We show that our dataset is very challenging, with all evaluated methods failing on its hardest parts.
As part of the dataset release, we provide the tooling used to generate it, enabling efficient and effective 2D correspondence annotation.
arXiv Detail & Related papers (2021-09-09T19:25:48Z) - Fog Simulation on Real LiDAR Point Clouds for 3D Object Detection in
Adverse Weather [92.84066576636914]
This work addresses the challenging task of LiDAR-based 3D object detection in foggy weather.
We tackle this problem by simulating physically accurate fog into clear-weather scenes.
We are the first to provide strong 3D object detection baselines on the Seeing Through Fog dataset.
arXiv Detail & Related papers (2021-08-11T14:37:54Z) - Lidar Light Scattering Augmentation (LISA): Physics-based Simulation of
Adverse Weather Conditions for 3D Object Detection [60.89616629421904]
Lidar-based object detectors are critical parts of the 3D perception pipeline in autonomous navigation systems such as self-driving cars.
They are sensitive to adverse weather conditions such as rain, snow and fog due to reduced signal-to-noise ratio (SNR) and signal-to-background ratio (SBR)
arXiv Detail & Related papers (2021-07-14T21:10:47Z) - A Flow Base Bi-path Network for Cross-scene Video Crowd Understanding in
Aerial View [93.23947591795897]
In this paper, we strive to tackle the challenges and automatically understand the crowd from the visual data collected from drones.
To alleviate the background noise generated in cross-scene testing, a double-stream crowd counting model is proposed.
To tackle the crowd density estimation problem under extreme dark environments, we introduce synthetic data generated by game Grand Theft Auto V(GTAV)
arXiv Detail & Related papers (2020-09-29T01:48:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.