LMT-Net: Lane Model Transformer Network for Automated HD Mapping from Sparse Vehicle Observations
- URL: http://arxiv.org/abs/2409.12409v1
- Date: Thu, 19 Sep 2024 02:14:35 GMT
- Title: LMT-Net: Lane Model Transformer Network for Automated HD Mapping from Sparse Vehicle Observations
- Authors: Michael Mink, Thomas Monninger, Steffen Staab,
- Abstract summary: Lane Model Transformer Network (LMT-Net) is an encoder-decoder neural network architecture that performs polyline encoding and predicts lane pairs and their connectivity.
We evaluate the performance of LMT-Net on an internal dataset that consists of multiple vehicle observations as well as human annotations as Ground Truth (GT)
- Score: 11.395749549636868
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In autonomous driving, High Definition (HD) maps provide a complete lane model that is not limited by sensor range and occlusions. However, the generation and upkeep of HD maps involves periodic data collection and human annotations, limiting scalability. To address this, we investigate automating the lane model generation and the use of sparse vehicle observations instead of dense sensor measurements. For our approach, a pre-processing step generates polylines by aligning and aggregating observed lane boundaries. Aligned driven traces are used as starting points for predicting lane pairs defined by the left and right boundary points. We propose Lane Model Transformer Network (LMT-Net), an encoder-decoder neural network architecture that performs polyline encoding and predicts lane pairs and their connectivity. A lane graph is formed by using predicted lane pairs as nodes and predicted lane connectivity as edges. We evaluate the performance of LMT-Net on an internal dataset that consists of multiple vehicle observations as well as human annotations as Ground Truth (GT). The evaluation shows promising results and demonstrates superior performance compared to the implemented baseline on both highway and non-highway Operational Design Domain (ODD).
Related papers
- Neural Semantic Map-Learning for Autonomous Vehicles [85.8425492858912]
We present a mapping system that fuses local submaps gathered from a fleet of vehicles at a central instance to produce a coherent map of the road environment.
Our method jointly aligns and merges the noisy and incomplete local submaps using a scene-specific Neural Signed Distance Field.
We leverage memory-efficient sparse feature-grids to scale to large areas and introduce a confidence score to model uncertainty in scene reconstruction.
arXiv Detail & Related papers (2024-10-10T10:10:03Z) - LaneSegNet: Map Learning with Lane Segment Perception for Autonomous
Driving [60.55208681215818]
We introduce LaneSegNet, the first end-to-end mapping network generating lane segments to obtain a complete representation of the road structure.
Our algorithm features two key modifications. One is a lane attention module to capture pivotal region details within the long-range feature space.
On the OpenLane-V2 dataset, LaneSegNet outperforms previous counterparts by a substantial gain across three tasks.
arXiv Detail & Related papers (2023-12-26T16:22:10Z) - G-MEMP: Gaze-Enhanced Multimodal Ego-Motion Prediction in Driving [71.9040410238973]
We focus on inferring the ego trajectory of a driver's vehicle using their gaze data.
Next, we develop G-MEMP, a novel multimodal ego-trajectory prediction network that combines GPS and video input with gaze data.
The results show that G-MEMP significantly outperforms state-of-the-art methods in both benchmarks.
arXiv Detail & Related papers (2023-12-13T23:06:30Z) - Prior Based Online Lane Graph Extraction from Single Onboard Camera
Image [133.68032636906133]
We tackle online estimation of the lane graph from a single onboard camera image.
The prior is extracted from the dataset through a transformer based Wasserstein Autoencoder.
The autoencoder is then used to enhance the initial lane graph estimates.
arXiv Detail & Related papers (2023-07-25T08:58:26Z) - Inverting the Fundamental Diagram and Forecasting Boundary Conditions:
How Machine Learning Can Improve Macroscopic Models for Traffic Flow [0.0]
We consider a dataset with flux and velocity data of vehicles moving on a highway, collected by fixed sensors and classified by lane and by class of vehicle.
We extrapolate two important pieces of information: 1) if congestion is appearing under the sensor, and 2) the total amount of vehicles which is going to pass under the sensor in the next future.
These pieces of information are then used to improve the accuracy of an LWR-based first-order multi-class model describing the dynamics of traffic flow between sensors.
arXiv Detail & Related papers (2023-03-21T11:07:19Z) - Fully End-to-end Autonomous Driving with Semantic Depth Cloud Mapping
and Multi-Agent [2.512827436728378]
We propose a novel deep learning model trained with end-to-end and multi-task learning manners to perform both perception and control tasks simultaneously.
The model is evaluated on CARLA simulator with various scenarios made of normal-adversarial situations and different weathers to mimic real-world conditions.
arXiv Detail & Related papers (2022-04-12T03:57:01Z) - Lane Graph Estimation for Scene Understanding in Urban Driving [34.82775302794312]
We propose a novel approach for lane geometry estimation from bird's-eye-view images.
We train a graph estimation model on multimodal bird's-eye-view data processed from the popular NuScenes dataset.
Our model shows promising performance for most evaluated urban scenes and can serve as a step towards automated generation of HD lane annotations for autonomous driving.
arXiv Detail & Related papers (2021-05-01T08:38:18Z) - DAGMapper: Learning to Map by Discovering Lane Topology [84.12949740822117]
We focus on drawing the lane boundaries of complex highways with many lanes that contain topology changes due to forks and merges.
We formulate the problem as inference in a directed acyclic graphical model (DAG), where the nodes of the graph encode geometric and topological properties of the local regions of the lane boundaries.
We show the effectiveness of our approach on two major North American Highways in two different states and show high precision and recall as well as 89% correct topology.
arXiv Detail & Related papers (2020-12-22T21:58:57Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - RONELD: Robust Neural Network Output Enhancement for Active Lane
Detection [1.3965477771846408]
Recent state-of-the-art lane detection algorithms utilize convolutional neural networks (CNNs) to train deep learning models.
We present a real-time robust neural network output enhancement for active lane detection (RONELD)
Experimental results demonstrate an up to two-fold increase in accuracy using RONELD.
arXiv Detail & Related papers (2020-10-19T14:22:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.