Detection-segmentation convolutional neural network for autonomous
vehicle perception
- URL: http://arxiv.org/abs/2306.17485v1
- Date: Fri, 30 Jun 2023 08:54:52 GMT
- Title: Detection-segmentation convolutional neural network for autonomous
vehicle perception
- Authors: Maciej Baczmanski, Robert Synoczek, Mateusz Wasala, Tomasz Kryjak
- Abstract summary: Object detection and segmentation are two core modules of an autonomous vehicle perception system.
Currently, the most commonly used algorithms are based on deep neural networks, which guarantee high efficiency but require high-performance computing platforms.
A reduction in the complexity of the network can be achieved by using an appropriate architecture, representation, and computing platform.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object detection and segmentation are two core modules of an autonomous
vehicle perception system. They should have high efficiency and low latency
while reducing computational complexity. Currently, the most commonly used
algorithms are based on deep neural networks, which guarantee high efficiency
but require high-performance computing platforms. In the case of autonomous
vehicles, i.e. cars, but also drones, it is necessary to use embedded platforms
with limited computing power, which makes it difficult to meet the requirements
described above. A reduction in the complexity of the network can be achieved
by using an appropriate: architecture, representation (reduced numerical
precision, quantisation, pruning), and computing platform. In this paper, we
focus on the first factor - the use of so-called detection-segmentation
networks as a component of a perception system. We considered the task of
segmenting the drivable area and road markings in combination with the
detection of selected objects (pedestrians, traffic lights, and obstacles). We
compared the performance of three different architectures described in the
literature: MultiTask V3, HybridNets, and YOLOP. We conducted the experiments
on a custom dataset consisting of approximately 500 images of the drivable area
and lane markings, and 250 images of detected objects. Of the three methods
analysed, MultiTask V3 proved to be the best, achieving 99% mAP_50 for
detection, 97% MIoU for drivable area segmentation, and 91% MIoU for lane
segmentation, as well as 124 fps on the RTX 3060 graphics card. This
architecture is a good solution for embedded perception systems for autonomous
vehicles. The code is available at: https://github.com/vision-agh/MMAR_2023.
Related papers
- Implementation of a perception system for autonomous vehicles using a
detection-segmentation network in SoC FPGA [0.0]
We have used the MultiTaskV3 detection-segmentation network as the basis for a perception system that can perform both functionalities within a single architecture.
The whole system consumes relatively little power compared to a CPU-based implementation.
It also achieves an accuracy higher than 97% of the mAP for object detection and above 90% of the mIoU for image segmentation.
arXiv Detail & Related papers (2023-07-17T17:44:18Z) - HeteroEdge: Addressing Asymmetry in Heterogeneous Collaborative
Autonomous Systems [1.274065448486689]
We propose a self-adaptive optimization framework for a testbed comprising two Unmanned Ground Vehicles (UGVs) and two NVIDIA Jetson devices.
This framework efficiently manages multiple tasks (storage, processing, computation, transmission, inference) on heterogeneous nodes concurrently.
It involves compressing and masking input image frames, identifying similar frames, and profiling devices to obtain boundary conditions for optimization.
arXiv Detail & Related papers (2023-05-05T02:43:16Z) - Architecturing Binarized Neural Networks for Traffic Sign Recognition [0.0]
Binarized neural networks (BNNs) have shown promising results in computationally limited and energy-constrained devices.
We propose BNNs architectures which achieve more than $90%$ for the German Traffic Sign Recognition Benchmark (GTSRB)
The number of parameters of these architectures varies from 100k to less than 2M.
arXiv Detail & Related papers (2023-03-27T08:46:31Z) - SPIN Road Mapper: Extracting Roads from Aerial Images via Spatial and
Interaction Space Graph Reasoning for Autonomous Driving [64.10636296274168]
Road extraction is an essential step in building autonomous navigation systems.
Using just convolution neural networks (ConvNets) for this problem is not effective as it is inefficient at capturing distant dependencies between road segments in the image.
We propose a Spatial and Interaction Space Graph Reasoning (SPIN) module which when plugged into a ConvNet performs reasoning over graphs constructed on spatial and interaction spaces projected from the feature maps.
arXiv Detail & Related papers (2021-09-16T03:52:17Z) - CFTrack: Center-based Radar and Camera Fusion for 3D Multi-Object
Tracking [9.62721286522053]
We propose an end-to-end network for joint object detection and tracking based on radar and camera sensor fusion.
Our proposed method uses a center-based radar-camera fusion algorithm for object detection and utilizes a greedy algorithm for object association.
We evaluate our method on the challenging nuScenes dataset, where it achieves 20.0 AMOTA and outperforms all vision-based 3D tracking methods in the benchmark.
arXiv Detail & Related papers (2021-07-11T23:56:53Z) - Dynamic Fusion Module Evolves Drivable Area and Road Anomaly Detection:
A Benchmark and Algorithms [16.417299198546168]
Joint detection of drivable areas and road anomalies is very important for mobile robots.
In this paper, we first build a drivable area and road anomaly detection benchmark for ground mobile robots.
We propose a novel module, referred to as the dynamic fusion module (DFM), which can be easily deployed in existing data-fusion networks.
arXiv Detail & Related papers (2021-03-03T14:38:27Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z) - Binary DAD-Net: Binarized Driveable Area Detection Network for
Autonomous Driving [94.40107679615618]
This paper proposes a novel binarized driveable area detection network (binary DAD-Net)
It uses only binary weights and activations in the encoder, the bottleneck, and the decoder part.
It outperforms state-of-the-art semantic segmentation networks on public datasets.
arXiv Detail & Related papers (2020-06-15T07:09:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.