Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across
Different Shutter Types
- URL: http://arxiv.org/abs/2309.08136v1
- Date: Fri, 15 Sep 2023 04:07:42 GMT
- Title: Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across
Different Shutter Types
- Authors: Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel
- Abstract summary: This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset.
In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets.
- Score: 7.0441427250832644
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Computer vision (CV) pipelines are typically evaluated on datasets processed
by image signal processing (ISP) pipelines even though, for
resource-constrained applications, an important research goal is to avoid as
many ISP steps as possible. In particular, most CV datasets consist of global
shutter (GS) images even though most cameras today use a rolling shutter (RS).
This paper studies the impact of different shutter mechanisms on machine
learning (ML) object detection models on a synthetic dataset that we generate
using the advanced simulation capabilities of Unreal Engine 5 (UE5). In
particular, we train and evaluate mainstream detection models with our
synthetically-generated paired GS and RS datasets to ascertain whether there
exists a significant difference in detection accuracy between these two shutter
modalities, especially when capturing low-speed objects (e.g., pedestrians).
The results of this emulation framework indicate the performance between them
are remarkably congruent for coarse-grained detection (mean average precision
(mAP) for IOU=0.5), but have significant differences for fine-grained measures
of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines
might not need explicit correction for RS for many object detection
applications, but mitigating RS effects in ISP-less ML pipelines that target
fine-grained location of the objects may need additional research.
Related papers
- OpticalRS-4M: Scaling Efficient Masked Autoencoder Learning on Large Remote Sensing Dataset [66.15872913664407]
We present a new pre-training pipeline for RS models, featuring the creation of a large-scale RS dataset and an efficient MIM approach.
We curated a high-quality dataset named OpticalRS-4M by collecting publicly available RS datasets and processing them through exclusion, slicing, and deduplication.
Experiments demonstrate that OpticalRS-4M significantly improves classification, detection, and segmentation performance, while SelectiveMAE increases training efficiency over 2 times.
arXiv Detail & Related papers (2024-06-17T15:41:57Z) - Efficient Meta-Learning Enabled Lightweight Multiscale Few-Shot Object Detection in Remote Sensing Images [15.12889076965307]
YOLOv7 one-stage detector is subjected to a novel meta-learning training framework.
This transformation allows the detector to adeptly address FSOD tasks while capitalizing on its inherent advantage of lightweight.
To validate the effectiveness of our proposed detector, we conducted performance comparisons with current state-of-the-art detectors.
arXiv Detail & Related papers (2024-04-29T04:56:52Z) - SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - ParGANDA: Making Synthetic Pedestrians A Reality For Object Detection [2.7648976108201815]
We propose to use a Generative Adversarial Network (GAN) to close the gap between the real and synthetic data.
Our approach not only produces visually plausible samples but also does not require any labels of the real domain.
arXiv Detail & Related papers (2023-07-21T05:26:32Z) - A Simulation-Augmented Benchmarking Framework for Automatic RSO Streak
Detection in Single-Frame Space Images [7.457841062817294]
Deep convolutional neural networks (DCNNs) have shown superior performance in object detection when large-scale datasets are available.
We introduce a novel simulation-augmented benchmarking framework for RSO detection (SAB-RSOD)
In our framework, by making the best use of the hardware parameters of the sensor that captures real-world space images, we first develop a high-fidelity RSO simulator.
Then, we use this simulator to generate images that contain diversified RSOs in space and annotate them automatically.
arXiv Detail & Related papers (2023-04-30T07:00:16Z) - Comprehensive Analysis of the Object Detection Pipeline on UAVs [16.071349046409885]
We first empirically analyze the influence of seven parameters (quantization, compression, resolution, color model, image distortion, gamma correction, additional channels) in remote sensing applications.
We show that not all parameters have an equal impact on detection accuracy and data throughput, and that by using a suitable compromise between parameters we are able to improve detection accuracy for lightweight object detection models.
arXiv Detail & Related papers (2022-03-01T09:30:01Z) - On the Role of Sensor Fusion for Object Detection in Future Vehicular
Networks [25.838878314196375]
We evaluate how using a combination of different sensors affects the detection of the environment in which the vehicles move and operate.
The final objective is to identify the optimal setup that would minimize the amount of data to be distributed over the channel.
arXiv Detail & Related papers (2021-04-23T18:58:37Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Learning Camera Miscalibration Detection [83.38916296044394]
This paper focuses on a data-driven approach to learn the detection of miscalibration in vision sensors, specifically RGB cameras.
Our contributions include a proposed miscalibration metric for RGB cameras and a novel semi-synthetic dataset generation pipeline based on this metric.
By training a deep convolutional neural network, we demonstrate the effectiveness of our pipeline to identify whether a recalibration of the camera's intrinsic parameters is required or not.
arXiv Detail & Related papers (2020-05-24T10:32:49Z) - A DICOM Framework for Machine Learning Pipelines against Real-Time
Radiology Images [50.222197963803644]
Niffler is an integrated framework that enables the execution of machine learning pipelines at research clusters.
Niffler uses the Digital Imaging and Communications in Medicine (DICOM) protocol to fetch and store imaging data.
We present its architecture and three of its use cases: an inferior vena cava filter detection from the images in real-time, identification of scanner utilization, and scanner clock calibration.
arXiv Detail & Related papers (2020-04-16T21:06:49Z) - Stance Detection Benchmark: How Robust Is Your Stance Detection? [65.91772010586605]
Stance Detection (StD) aims to detect an author's stance towards a certain topic or claim.
We introduce a StD benchmark that learns from ten StD datasets of various domains in a multi-dataset learning setting.
Within this benchmark setup, we are able to present new state-of-the-art results on five of the datasets.
arXiv Detail & Related papers (2020-01-06T13:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.