Graph-based Topology Reasoning for Driving Scenes
- URL: http://arxiv.org/abs/2304.05277v2
- Date: Mon, 28 Aug 2023 10:51:09 GMT
- Title: Graph-based Topology Reasoning for Driving Scenes
- Authors: Tianyu Li, Li Chen, Huijie Wang, Yang Li, Jiazhi Yang, Xiangwei Geng,
Shengyin Jiang, Yuting Wang, Hang Xu, Chunjing Xu, Junchi Yan, Ping Luo,
Hongyang Li
- Abstract summary: We present TopoNet, the first end-to-end framework capable of abstracting traffic knowledge beyond conventional perception tasks.
We evaluate TopoNet on the challenging scene understanding benchmark, OpenLane-V2.
- Score: 102.35885039110057
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Understanding the road genome is essential to realize autonomous driving.
This highly intelligent problem contains two aspects - the connection
relationship of lanes, and the assignment relationship between lanes and
traffic elements, where a comprehensive topology reasoning method is vacant. On
one hand, previous map learning techniques struggle in deriving lane
connectivity with segmentation or laneline paradigms; or prior lane
topology-oriented approaches focus on centerline detection and neglect the
interaction modeling. On the other hand, the traffic element to lane assignment
problem is limited in the image domain, leaving how to construct the
correspondence from two views an unexplored challenge. To address these issues,
we present TopoNet, the first end-to-end framework capable of abstracting
traffic knowledge beyond conventional perception tasks. To capture the driving
scene topology, we introduce three key designs: (1) an embedding module to
incorporate semantic knowledge from 2D elements into a unified feature space;
(2) a curated scene graph neural network to model relationships and enable
feature interaction inside the network; (3) instead of transmitting messages
arbitrarily, a scene knowledge graph is devised to differentiate prior
knowledge from various types of the road genome. We evaluate TopoNet on the
challenging scene understanding benchmark, OpenLane-V2, where our approach
outperforms all previous works by a great margin on all perceptual and
topological metrics. The code is released at
https://github.com/OpenDriveLab/TopoNet
Related papers
- LaneSegNet: Map Learning with Lane Segment Perception for Autonomous
Driving [60.55208681215818]
We introduce LaneSegNet, the first end-to-end mapping network generating lane segments to obtain a complete representation of the road structure.
Our algorithm features two key modifications. One is a lane attention module to capture pivotal region details within the long-range feature space.
On the OpenLane-V2 dataset, LaneSegNet outperforms previous counterparts by a substantial gain across three tasks.
arXiv Detail & Related papers (2023-12-26T16:22:10Z) - TopoMask: Instance-Mask-Based Formulation for the Road Topology Problem
via Transformer-Based Architecture [4.970364068620607]
We introduce TopoMask for predicting centerlines in road topology.
TopoMask has ranked 4th in the OpenLane-V2 Score (OLS) and 2nd in the F1 score of centerline prediction in OpenLane Topology Challenge 2023.
arXiv Detail & Related papers (2023-06-08T17:58:57Z) - How To Not Train Your Dragon: Training-free Embodied Object Goal
Navigation with Semantic Frontiers [94.46825166907831]
We present a training-free solution to tackle the object goal navigation problem in Embodied AI.
Our method builds a structured scene representation based on the classic visual simultaneous localization and mapping (V-SLAM) framework.
Our method propagates semantics on the scene graphs based on language priors and scene statistics to introduce semantic knowledge to the geometric frontiers.
arXiv Detail & Related papers (2023-05-26T13:38:33Z) - OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping [84.65114565766596]
We present OpenLane-V2, the first dataset on topology reasoning for traffic scene structure.
OpenLane-V2 consists of 2,000 annotated road scenes that describe traffic elements and their correlation to the lanes.
We evaluate various state-of-the-art methods, and present their quantitative and qualitative results on OpenLane-V2 to indicate future avenues for investigating topology reasoning in traffic scenes.
arXiv Detail & Related papers (2023-04-20T16:31:22Z) - JPerceiver: Joint Perception Network for Depth, Pose and Layout
Estimation in Driving Scenes [75.20435924081585]
JPerceiver can simultaneously estimate scale-aware depth and VO as well as BEV layout from a monocular video sequence.
It exploits the cross-view geometric transformation (CGT) to propagate the absolute scale from the road layout to depth and VO.
Experiments on Argoverse, Nuscenes and KITTI show the superiority of JPerceiver over existing methods on all the above three tasks.
arXiv Detail & Related papers (2022-07-16T10:33:59Z) - Autonomous Navigation through intersections with Graph
ConvolutionalNetworks and Conditional Imitation Learning for Self-driving
Cars [10.080958939027363]
In autonomous driving, navigation through unsignaled intersections is a challenging task.
We propose a novel branched network G-CIL for the navigation policy learning.
Our end-to-end trainable neural network outperforms the baselines with higher success rate and shorter navigation time.
arXiv Detail & Related papers (2021-02-01T07:33:12Z) - DAGMapper: Learning to Map by Discovering Lane Topology [84.12949740822117]
We focus on drawing the lane boundaries of complex highways with many lanes that contain topology changes due to forks and merges.
We formulate the problem as inference in a directed acyclic graphical model (DAG), where the nodes of the graph encode geometric and topological properties of the local regions of the lane boundaries.
We show the effectiveness of our approach on two major North American Highways in two different states and show high precision and recall as well as 89% correct topology.
arXiv Detail & Related papers (2020-12-22T21:58:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.