LDNet: End-to-End Lane Marking Detection Approach Using a Dynamic Vision
Sensor
- URL: http://arxiv.org/abs/2009.08020v2
- Date: Tue, 30 Nov 2021 10:08:40 GMT
- Title: LDNet: End-to-End Lane Marking Detection Approach Using a Dynamic Vision
Sensor
- Authors: Farzeen Munir (Student Member, IEEE), Shoaib Azam (Student Member,
IEEE), Moongu Jeon (Senior Member, IEEE), Byung-Geun Lee (Member, IEEE), and
Witold Pedrycz (Life Fellow, IEEE)
- Abstract summary: This paper explores the novel application of lane marking detection using an event camera.
The spatial resolution of the encoded features is retained by a dense atrous spatial pyramid pooling block.
The efficacy of the proposed work is evaluated using the DVS dataset for lane extraction.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern vehicles are equipped with various driver-assistance systems,
including automatic lane keeping, which prevents unintended lane departures.
Traditional lane detection methods incorporate handcrafted or deep
learning-based features followed by postprocessing techniques for lane
extraction using frame-based RGB cameras. The utilization of frame-based RGB
cameras for lane detection tasks is prone to illumination variations, sun
glare, and motion blur, which limits the performance of lane detection methods.
Incorporating an event camera for lane detection tasks in the perception stack
of autonomous driving is one of the most promising solutions for mitigating
challenges encountered by frame-based RGB cameras. The main contribution of
this work is the design of the lane marking detection model, which employs the
dynamic vision sensor. This paper explores the novel application of lane
marking detection using an event camera by designing a convolutional encoder
followed by the attention-guided decoder. The spatial resolution of the encoded
features is retained by a dense atrous spatial pyramid pooling (ASPP) block.
The additive attention mechanism in the decoder improves performance for high
dimensional input encoded features that promote lane localization and relieve
postprocessing computation. The efficacy of the proposed work is evaluated
using the DVS dataset for lane extraction (DET). The experimental results show
a significant improvement of $5.54\%$ and $5.03\%$ in $F1$ scores in multiclass
and binary-class lane marking detection tasks. Additionally, the intersection
over union ($IoU$) scores of the proposed method surpass those of the
best-performing state-of-the-art method by $6.50\%$ and $9.37\%$ in multiclass
and binary-class tasks, respectively.
Related papers
- Monocular Lane Detection Based on Deep Learning: A Survey [51.19079381823076]
Lane detection plays an important role in autonomous driving perception systems.
As deep learning algorithms gain popularity, monocular lane detection methods based on deep learning have demonstrated superior performance.
This paper presents a comprehensive overview of existing methods, encompassing both the increasingly mature 2D lane detection approaches and the developing 3D lane detection works.
arXiv Detail & Related papers (2024-11-25T12:09:43Z) - Multi-Modal 3D Object Detection by Box Matching [109.43430123791684]
We propose a novel Fusion network by Box Matching (FBMNet) for multi-modal 3D detection.
With the learned assignments between 3D and 2D object proposals, the fusion for detection can be effectively performed by combing their ROI features.
arXiv Detail & Related papers (2023-05-12T18:08:51Z) - Laneformer: Object-aware Row-Column Transformers for Lane Detection [96.62919884511287]
Laneformer is a transformer-based architecture tailored for lane detection in autonomous driving.
Inspired by recent advances of the transformer encoder-decoder architecture in various vision tasks, we move forwards to design a new end-to-end Laneformer architecture.
arXiv Detail & Related papers (2022-03-18T10:14:35Z) - A Pedestrian Detection and Tracking Framework for Autonomous Cars:
Efficient Fusion of Camera and LiDAR Data [0.17205106391379021]
This paper presents a novel method for pedestrian detection and tracking by fusing camera and LiDAR sensor data.
The detection phase is performed by converting LiDAR streams to computationally tractable depth images, and then, a deep neural network is developed to identify pedestrian candidates.
The tracking phase is a combination of the Kalman filter prediction and an optical flow algorithm to track multiple pedestrians in a scene.
arXiv Detail & Related papers (2021-08-27T16:16:01Z) - CFTrack: Center-based Radar and Camera Fusion for 3D Multi-Object
Tracking [9.62721286522053]
We propose an end-to-end network for joint object detection and tracking based on radar and camera sensor fusion.
Our proposed method uses a center-based radar-camera fusion algorithm for object detection and utilizes a greedy algorithm for object association.
We evaluate our method on the challenging nuScenes dataset, where it achieves 20.0 AMOTA and outperforms all vision-based 3D tracking methods in the benchmark.
arXiv Detail & Related papers (2021-07-11T23:56:53Z) - Preprocessing Methods of Lane Detection and Tracking for Autonomous
Driving [0.0]
Real time lane detection and tracking (LDT) is one of the most consequential parts to performing the above tasks.
In this paper, we survey preprocessing methods for detecting lane marking as well as tracking lane boundaries in real time focusing on vision-based system.
arXiv Detail & Related papers (2021-04-10T13:03:52Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Structure-Aware Network for Lane Marker Extraction with Dynamic Vision
Sensor [14.55881454495574]
We introduce Dynamic Vision Sensor (DVS), a type of event-based sensor to lane marker extraction task.
We generate high-resolution DVS dataset for lane marker extraction with resolution of 1280$times$800 pixels.
We then propose a structure-aware network for lane marker extraction in DVS images.
We evaluate our proposed network with other state-of-the-art lane marker extraction models on this dataset.
arXiv Detail & Related papers (2020-08-14T06:28:20Z) - Heatmap-based Vanishing Point boosts Lane Detection [3.8170259685864165]
We propose a new multi-task fusion network architecture for high-precision lane detection.
The proposed fusion strategy was tested using the public CULane dataset.
The experimental results suggest that the lane detection accuracy of our method outperforms those of state-of-the-art (SOTA) methods.
arXiv Detail & Related papers (2020-07-30T17:17:00Z) - Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning [59.19469551774703]
Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
arXiv Detail & Related papers (2020-03-05T05:29:44Z) - Road Curb Detection and Localization with Monocular Forward-view Vehicle
Camera [74.45649274085447]
We propose a robust method for estimating road curb 3D parameters using a calibrated monocular camera equipped with a fisheye lens.
Our approach is able to estimate the vehicle to curb distance in real time with mean accuracy of more than 90%.
arXiv Detail & Related papers (2020-02-28T00:24:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.