ONCE-3DLanes: Building Monocular 3D Lane Detection
- URL: http://arxiv.org/abs/2205.00301v1
- Date: Sat, 30 Apr 2022 16:35:25 GMT
- Title: ONCE-3DLanes: Building Monocular 3D Lane Detection
- Authors: Fan Yan, Ming Nie, Xinyue Cai, Jianhua Han, Hang Xu, Zhen Yang,
Chaoqiang Ye, Yanwei Fu, Michael Bi Mi, Li Zhang
- Abstract summary: We present ONCE-3DLanes, a real-world autonomous driving dataset with lane layout annotation in 3D space.
By exploiting the explicit relationship between point clouds and image pixels, a dataset annotation pipeline is designed to automatically generate high-quality 3D lane locations.
- Score: 41.46466150783367
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present ONCE-3DLanes, a real-world autonomous driving dataset with lane
layout annotation in 3D space. Conventional 2D lane detection from a monocular
image yields poor performance of following planning and control tasks in
autonomous driving due to the case of uneven road. Predicting the 3D lane
layout is thus necessary and enables effective and safe driving. However,
existing 3D lane detection datasets are either unpublished or synthesized from
a simulated environment, severely hampering the development of this field. In
this paper, we take steps towards addressing these issues. By exploiting the
explicit relationship between point clouds and image pixels, a dataset
annotation pipeline is designed to automatically generate high-quality 3D lane
locations from 2D lane annotations in 211K road scenes. In addition, we present
an extrinsic-free, anchor-free method, called SALAD, regressing the 3D
coordinates of lanes in image view without converting the feature map into the
bird's-eye view (BEV). To facilitate future research on 3D lane detection, we
benchmark the dataset and provide a novel evaluation metric, performing
extensive experiments of both existing approaches and our proposed method. The
aim of our work is to revive the interest of 3D lane detection in a real-world
scenario. We believe our work can lead to the expected and unexpected
innovations in both academia and industry.
Related papers
- HeightFormer: A Semantic Alignment Monocular 3D Object Detection Method from Roadside Perspective [11.841338298700421]
We propose a novel 3D object detection framework integrating Spatial Former and Voxel Pooling Former to enhance 2D-to-3D projection based on height estimation.
Experiments were conducted using the Rope3D and DAIR-V2X-I dataset, and the results demonstrated the outperformance of the proposed algorithm in the detection of both vehicles and cyclists.
arXiv Detail & Related papers (2024-10-10T09:37:33Z) - Enhancing 3D Lane Detection and Topology Reasoning with 2D Lane Priors [40.92232275558338]
3D lane detection and topology reasoning are essential tasks in autonomous driving scenarios.
We propose Topo2D, a novel framework based on Transformer, leveraging 2D lane instances to initialize 3D queries and 3D positional embeddings.
Topo2D achieves 44.5% OLS on multi-view topology reasoning benchmark OpenLane-V2 and 62.6% F-Socre on single-view 3D lane detection benchmark OpenLane.
arXiv Detail & Related papers (2024-06-05T09:48:56Z) - 3D Lane Detection from Front or Surround-View using Joint-Modeling & Matching [27.588395086563978]
We propose a joint modeling approach that combines Bezier curves and methods.
We also introduce a novel 3D Spatial, representing an exploration of 3D surround-view lane detection research.
This innovative method establishes a new benchmark in front-view 3D lane detection on the Openlane dataset.
arXiv Detail & Related papers (2024-01-16T01:12:24Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data [80.14669385741202]
We propose a self-supervised pre-training method for 3D perception models tailored to autonomous driving data.
We leverage the availability of synchronized and calibrated image and Lidar sensors in autonomous driving setups.
Our method does not require any point cloud nor image annotations.
arXiv Detail & Related papers (2022-03-30T12:40:30Z) - PersFormer: 3D Lane Detection via Perspective Transformer and the
OpenLane Benchmark [109.03773439461615]
PersFormer is an end-to-end monocular 3D lane detector with a novel Transformer-based spatial feature transformation module.
We release one of the first large-scale real-world 3D lane datasets, called OpenLane, with high-quality annotation and scenario diversity.
arXiv Detail & Related papers (2022-03-21T16:12:53Z) - FGR: Frustum-Aware Geometric Reasoning for Weakly Supervised 3D Vehicle
Detection [81.79171905308827]
We propose frustum-aware geometric reasoning (FGR) to detect vehicles in point clouds without any 3D annotations.
Our method consists of two stages: coarse 3D segmentation and 3D bounding box estimation.
It is able to accurately detect objects in 3D space with only 2D bounding boxes and sparse point clouds.
arXiv Detail & Related papers (2021-05-17T07:29:55Z) - Monocular 3D Vehicle Detection Using Uncalibrated Traffic Cameras
through Homography [12.062095895630563]
This paper proposes a method to extract the position and pose of vehicles in the 3D world from a single traffic camera.
We observe that the homography between the road plane and the image plane is essential to 3D vehicle detection.
We propose a new regression target called textittailedr-box and a textitdual-view network architecture which boosts the detection accuracy on warped BEV images.
arXiv Detail & Related papers (2021-03-29T02:57:37Z) - Road Curb Detection and Localization with Monocular Forward-view Vehicle
Camera [74.45649274085447]
We propose a robust method for estimating road curb 3D parameters using a calibrated monocular camera equipped with a fisheye lens.
Our approach is able to estimate the vehicle to curb distance in real time with mean accuracy of more than 90%.
arXiv Detail & Related papers (2020-02-28T00:24:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.