Traffic Sign Detection With Event Cameras and DCNN
- URL: http://arxiv.org/abs/2207.13345v1
- Date: Wed, 27 Jul 2022 08:01:54 GMT
- Title: Traffic Sign Detection With Event Cameras and DCNN
- Authors: Piotr Wzorek and Tomasz Kryjak
- Abstract summary: Event cameras (DVS) have been used in vision systems as an alternative or supplement to traditional cameras.
In this work, we test whether these rather novel sensors can be applied to the popular task of traffic sign detection.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, event cameras (DVS - Dynamic Vision Sensors) have been used
in vision systems as an alternative or supplement to traditional cameras. They
are characterised by high dynamic range, high temporal resolution, low latency,
and reliable performance in limited lighting conditions -- parameters that are
particularly important in the context of advanced driver assistance systems
(ADAS) and self-driving cars. In this work, we test whether these rather novel
sensors can be applied to the popular task of traffic sign detection. To this
end, we analyse different representations of the event data: event frame, event
frequency, and the exponentially decaying time surface, and apply video frame
reconstruction using a deep neural network called FireNet. We use the deep
convolutional neural network YOLOv4 as a detector. For particular
representations, we obtain a detection accuracy in the range of 86.9-88.9%
mAP@0.5. The use of a fusion of the considered representations allows us to
obtain a detector with higher accuracy of 89.9% mAP@0.5. In comparison, the
detector for the frames reconstructed with FireNet is characterised by an
accuracy of 72.67% mAP@0.5. The results obtained illustrate the potential of
event cameras in automotive applications, either as standalone sensors or in
close cooperation with typical frame-based cameras.
Related papers
- Resource-Efficient Gesture Recognition using Low-Resolution Thermal
Camera via Spiking Neural Networks and Sparse Segmentation [1.7758299835471887]
This work proposes a novel approach for hand gesture recognition using an inexpensive, low-resolution (24 x 32) thermal sensor.
Compared to the use of standard RGB cameras, the proposed system is insensitive to lighting variations.
This paper shows that the innovative use of the recently proposed Monostable Multivibrator (MMV) neural networks as a new class of SNN achieves more than one order of magnitude smaller memory and compute complexity.
arXiv Detail & Related papers (2024-01-12T13:20:01Z) - Multi-Modal Neural Radiance Field for Monocular Dense SLAM with a
Light-Weight ToF Sensor [58.305341034419136]
We present the first dense SLAM system with a monocular camera and a light-weight ToF sensor.
We propose a multi-modal implicit scene representation that supports rendering both the signals from the RGB camera and light-weight ToF sensor.
Experiments demonstrate that our system well exploits the signals of light-weight ToF sensors and achieves competitive results.
arXiv Detail & Related papers (2023-08-28T07:56:13Z) - EventTransAct: A video transformer-based framework for Event-camera
based action recognition [52.537021302246664]
Event cameras offer new opportunities compared to standard action recognition in RGB videos.
In this study, we employ a computationally efficient model, namely the video transformer network (VTN), which initially acquires spatial embeddings per event-frame.
In order to better adopt the VTN for the sparse and fine-grained nature of event data, we design Event-Contrastive Loss ($mathcalL_EC$) and event-specific augmentations.
arXiv Detail & Related papers (2023-08-25T23:51:07Z) - Neuromorphic Seatbelt State Detection for In-Cabin Monitoring with Event
Cameras [0.932065750652415]
This research provides a proof of concept to expand event-based DMS techniques to include seatbelt state detection.
In a binary classification task, the fastened/unfastened frames were identified with an F1 score of 0.989 and 0.944 on the simulated and real test sets respectively.
arXiv Detail & Related papers (2023-08-15T14:27:46Z) - A Quality Index Metric and Method for Online Self-Assessment of
Autonomous Vehicles Sensory Perception [164.93739293097605]
We propose a novel evaluation metric, named as the detection quality index (DQI), which assesses the performance of camera-based object detection algorithms.
We have developed a superpixel-based attention network (SPA-NET) that utilizes raw image pixels and superpixels as input to predict the proposed DQI evaluation metric.
arXiv Detail & Related papers (2022-03-04T22:16:50Z) - ESL: Event-based Structured Light [62.77144631509817]
Event cameras are bio-inspired sensors providing significant advantages over standard cameras.
We propose a novel structured-light system using an event camera to tackle the problem of accurate and high-speed depth sensing.
arXiv Detail & Related papers (2021-11-30T15:47:39Z) - Event Guided Depth Sensing [50.997474285910734]
We present an efficient bio-inspired event-camera-driven depth estimation algorithm.
In our approach, we illuminate areas of interest densely, depending on the scene activity detected by the event camera.
We show the feasibility of our approach in a simulated autonomous driving sequences and real indoor environments.
arXiv Detail & Related papers (2021-10-20T11:41:11Z) - Fusion-FlowNet: Energy-Efficient Optical Flow Estimation using Sensor
Fusion and Deep Fused Spiking-Analog Network Architectures [7.565038387344594]
We present a sensor fusion framework for energy-efficient optical flow estimation using both frame- and event-based sensors.
Our network is end-to-end trained using unsupervised learning to avoid expensive video annotations.
arXiv Detail & Related papers (2021-03-19T02:03:33Z) - Combining Events and Frames using Recurrent Asynchronous Multimodal
Networks for Monocular Depth Prediction [51.072733683919246]
We introduce Recurrent Asynchronous Multimodal (RAM) networks to handle asynchronous and irregular data from multiple sensors.
Inspired by traditional RNNs, RAM networks maintain a hidden state that is updated asynchronously and can be queried at any time to generate a prediction.
We show an improvement over state-of-the-art methods by up to 30% in terms of mean depth absolute error.
arXiv Detail & Related papers (2021-02-18T13:24:35Z) - RODNet: A Real-Time Radar Object Detection Network Cross-Supervised by
Camera-Radar Fused Object 3D Localization [30.42848269877982]
We propose a deep radar object detection network, named RODNet, which is cross-supervised by a camera-radar fused algorithm.
Our proposed RODNet takes a sequence of RF images as the input to predict the likelihood of objects in the radar field of view (FoV)
With intensive experiments, our proposed cross-supervised RODNet achieves 86% average precision and 88% average recall of object detection performance.
arXiv Detail & Related papers (2021-02-09T22:01:55Z) - Real-Time Face & Eye Tracking and Blink Detection using Event Cameras [3.842206880015537]
Event cameras contain emerging, neuromorphic vision sensors that capture local light intensity changes at each pixel, generating a stream of asynchronous events.
Driver monitoring systems (DMS) are in-cabin safety systems designed to sense and understand a drivers physical and cognitive state.
This paper proposes a novel method to simultaneously detect and track faces and eyes for driver monitoring.
arXiv Detail & Related papers (2020-10-16T10:02:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.