End-to-end Lane Shape Prediction with Transformers
- URL: http://arxiv.org/abs/2011.04233v2
- Date: Sat, 28 Nov 2020 10:55:44 GMT
- Title: End-to-end Lane Shape Prediction with Transformers
- Authors: Ruijin Liu, Zejian Yuan, Tie Liu, Zhiliang Xiong
- Abstract summary: Lane detection is widely used for lane departure warning and adaptive cruise control in autonomous vehicles.
We propose an end-to-end method that directly outputs parameters of a lane shape model.
The proposed method is validated on the TuSimple benchmark and shows state-of-the-art accuracy with the most lightweight model size and fastest speed.
- Score: 13.103463647059634
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lane detection, the process of identifying lane markings as approximated
curves, is widely used for lane departure warning and adaptive cruise control
in autonomous vehicles. The popular pipeline that solves it in two steps --
feature extraction plus post-processing, while useful, is too inefficient and
flawed in learning the global context and lanes' long and thin structures. To
tackle these issues, we propose an end-to-end method that directly outputs
parameters of a lane shape model, using a network built with a transformer to
learn richer structures and context. The lane shape model is formulated based
on road structures and camera pose, providing physical interpretation for
parameters of network output. The transformer models non-local interactions
with a self-attention mechanism to capture slender structures and global
context. The proposed method is validated on the TuSimple benchmark and shows
state-of-the-art accuracy with the most lightweight model size and fastest
speed. Additionally, our method shows excellent adaptability to a challenging
self-collected lane detection dataset, showing its powerful deployment
potential in real applications. Codes are available at
https://github.com/liuruijin17/LSTR.
Related papers
- Homography Guided Temporal Fusion for Road Line and Marking Segmentation [73.47092021519245]
Road lines and markings are frequently occluded in the presence of moving vehicles, shadow, and glare.
We propose a Homography Guided Fusion (HomoFusion) module to exploit temporally-adjacent video frames for complementary cues.
We show that exploiting available camera intrinsic data and ground plane assumption for cross-frame correspondence can lead to a light-weight network with significantly improved performances in speed and accuracy.
arXiv Detail & Related papers (2024-04-11T10:26:40Z) - LDTR: Transformer-based Lane Detection with Anchor-chain Representation [11.184960972042406]
Lane detection scenarios with limited- or no-visual-clue of lanes remain challenging and crucial for automated driving.
Inspired by the DETR architecture, we propose LDTR, a transformer-based model to address these issues.
Experimental results demonstrate that LDTR achieves state-of-the-art performance on well-known datasets.
arXiv Detail & Related papers (2024-03-21T12:29:26Z) - Sketch and Refine: Towards Fast and Accurate Lane Detection [69.63287721343907]
Lane detection is a challenging task due to the complexity of real-world scenarios.
Existing approaches, whether proposal-based or keypoint-based, suffer from depicting lanes effectively and efficiently.
We present a "Sketch-and-Refine" paradigm that utilizes the merits of both keypoint-based and proposal-based methods.
Experiments show that our SRLane can run at a fast speed (i.e., 278 FPS) while yielding an F1 score of 78.9%.
arXiv Detail & Related papers (2024-01-26T09:28:14Z) - Decoupling the Curve Modeling and Pavement Regression for Lane Detection [67.22629246312283]
curve-based lane representation is a popular approach in many lane detection methods.
We propose a new approach to the lane detection task by decomposing it into two parts: curve modeling and ground height regression.
arXiv Detail & Related papers (2023-09-19T11:24:14Z) - HoughLaneNet: Lane Detection with Deep Hough Transform and Dynamic
Convolution [8.97991745734826]
Lanes can present difficulties for detection, as they can be narrow, fragmented, and often obscured by heavy traffic.
We propose a hierarchical Deep Hough Transform (DHT) approach that combines all lane features in an image into the Hough parameter space.
Our proposed network structure demonstrates improved performance in detecting heavily occluded or worn lane images.
arXiv Detail & Related papers (2023-07-07T10:08:29Z) - BSNet: Lane Detection via Draw B-spline Curves Nearby [21.40607319558899]
We revisit the curve-based lane detection methods from the perspectives of the lane representations' globality and locality.
We design a simple yet efficient network BSNet to ensure the acquisition of global and local features.
The proposed methods achieve state-of-the-art performance on the Tusimple, CULane, and LLAMAS datasets.
arXiv Detail & Related papers (2023-01-17T14:25:40Z) - Laneformer: Object-aware Row-Column Transformers for Lane Detection [96.62919884511287]
Laneformer is a transformer-based architecture tailored for lane detection in autonomous driving.
Inspired by recent advances of the transformer encoder-decoder architecture in various vision tasks, we move forwards to design a new end-to-end Laneformer architecture.
arXiv Detail & Related papers (2022-03-18T10:14:35Z) - Lane Detection with Versatile AtrousFormer and Local Semantic Guidance [92.83267435275802]
Lane detection is one of the core functions in autonomous driving.
Most existing methods tend to resort to CNN-based techniques.
We propose Atrous Transformer (AtrousFormer) to solve the problem.
arXiv Detail & Related papers (2022-03-08T13:25:35Z) - Rethinking Efficient Lane Detection via Curve Modeling [37.45243848960598]
The proposed method achieves a new state-of-the-art performance on the popular LLAMAS benchmark.
It also achieves favorable accuracy on the TuSimple and CU datasets, while retaining both low latency (> 150 FPS) and small model size ( 10M)
arXiv Detail & Related papers (2022-03-04T17:00:33Z) - CurveLane-NAS: Unifying Lane-Sensitive Architecture Search and Adaptive
Point Blending [102.98909328368481]
CurveLane-NAS is a novel lane-sensitive architecture search framework.
It captures both long-ranged coherent and accurate short-range curve information.
It unifies both architecture search and post-processing on curve lane predictions via point blending.
arXiv Detail & Related papers (2020-07-23T17:23:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.