RESA: Recurrent Feature-Shift Aggregator for Lane Detection
- URL: http://arxiv.org/abs/2008.13719v2
- Date: Thu, 25 Mar 2021 17:14:56 GMT
- Title: RESA: Recurrent Feature-Shift Aggregator for Lane Detection
- Authors: Tu Zheng, Hao Fang, Yi Zhang, Wenjian Tang, Zheng Yang, Haifeng Liu,
Deng Cai
- Abstract summary: We present a novel module named REcurrent Feature-Shift Aggregator (RESA) to enrich lane feature after preliminary feature extraction with an ordinary CNN.
RESA can conjecture lanes accurately in challenging scenarios with weak appearance clues by aggregating sliced feature map.
Our method achieves state-of-the-art results on two popular lane detection benchmarks (CULane and Tusimple)
- Score: 32.246537653680484
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lane detection is one of the most important tasks in self-driving. Due to
various complex scenarios (e.g., severe occlusion, ambiguous lanes, etc.) and
the sparse supervisory signals inherent in lane annotations, lane detection
task is still challenging. Thus, it is difficult for the ordinary convolutional
neural network (CNN) to train in general scenes to catch subtle lane feature
from the raw image. In this paper, we present a novel module named REcurrent
Feature-Shift Aggregator (RESA) to enrich lane feature after preliminary
feature extraction with an ordinary CNN. RESA takes advantage of strong shape
priors of lanes and captures spatial relationships of pixels across rows and
columns. It shifts sliced feature map recurrently in vertical and horizontal
directions and enables each pixel to gather global information. RESA can
conjecture lanes accurately in challenging scenarios with weak appearance clues
by aggregating sliced feature map. Moreover, we propose a Bilateral Up-Sampling
Decoder that combines coarse-grained and fine-detailed features in the
up-sampling stage. It can recover the low-resolution feature map into
pixel-wise prediction meticulously. Our method achieves state-of-the-art
results on two popular lane detection benchmarks (CULane and Tusimple). Code
has been made available at: https://github.com/ZJULearning/resa.
Related papers
- Deep Homography Estimation for Visual Place Recognition [49.235432979736395]
We propose a transformer-based deep homography estimation (DHE) network.
It takes the dense feature map extracted by a backbone network as input and fits homography for fast and learnable geometric verification.
Experiments on benchmark datasets show that our method can outperform several state-of-the-art methods.
arXiv Detail & Related papers (2024-02-25T13:22:17Z) - Sketch and Refine: Towards Fast and Accurate Lane Detection [69.63287721343907]
Lane detection is a challenging task due to the complexity of real-world scenarios.
Existing approaches, whether proposal-based or keypoint-based, suffer from depicting lanes effectively and efficiently.
We present a "Sketch-and-Refine" paradigm that utilizes the merits of both keypoint-based and proposal-based methods.
Experiments show that our SRLane can run at a fast speed (i.e., 278 FPS) while yielding an F1 score of 78.9%.
arXiv Detail & Related papers (2024-01-26T09:28:14Z) - RCLane: Relay Chain Prediction for Lane Detection [76.62424079494285]
We present a new method for lane detection based on relay chain prediction.
Our strategy allows us to establish new state-of-the-art on four major benchmarks including TuSimple, CULane, CurveLanes and LLAMAS.
arXiv Detail & Related papers (2022-07-19T16:48:39Z) - Ultra Fast Deep Lane Detection with Hybrid Anchor Driven Ordinal
Classification [13.735482211178931]
We propose a novel, simple, yet effective formulation aiming at ultra fast speed and the problem of challenging scenarios.
Specifically, we treat the process of lane detection as an anchor-driven ordinal classification problem using global features.
Our method could achieve state-of-the-art performance in terms of both speed and accuracy.
arXiv Detail & Related papers (2022-06-15T08:53:02Z) - Lane detection with Position Embedding [14.720302401952157]
We present a novel module to enrich lane feature after preliminary feature extraction with an ordinary CNN.
On the basis of RESA, we introduce the method of position embedding to enhance the spatial features.
The experimental results show that this method has achieved the best accuracy 96.93% on Tusimple dataset.
arXiv Detail & Related papers (2022-03-23T09:48:59Z) - PersFormer: 3D Lane Detection via Perspective Transformer and the
OpenLane Benchmark [109.03773439461615]
PersFormer is an end-to-end monocular 3D lane detector with a novel Transformer-based spatial feature transformation module.
We release one of the first large-scale real-world 3D lane datasets, called OpenLane, with high-quality annotation and scenario diversity.
arXiv Detail & Related papers (2022-03-21T16:12:53Z) - Lane Detection with Versatile AtrousFormer and Local Semantic Guidance [92.83267435275802]
Lane detection is one of the core functions in autonomous driving.
Most existing methods tend to resort to CNN-based techniques.
We propose Atrous Transformer (AtrousFormer) to solve the problem.
arXiv Detail & Related papers (2022-03-08T13:25:35Z) - Heatmap-based Vanishing Point boosts Lane Detection [3.8170259685864165]
We propose a new multi-task fusion network architecture for high-precision lane detection.
The proposed fusion strategy was tested using the public CULane dataset.
The experimental results suggest that the lane detection accuracy of our method outperforms those of state-of-the-art (SOTA) methods.
arXiv Detail & Related papers (2020-07-30T17:17:00Z) - Multiview Detection with Feature Perspective Transformation [59.34619548026885]
We propose a novel multiview detection system, MVDet.
We take an anchor-free approach to aggregate multiview information by projecting feature maps onto the ground plane.
Our entire model is end-to-end learnable and achieves 88.2% MODA on the standard Wildtrack dataset.
arXiv Detail & Related papers (2020-07-14T17:58:30Z) - LR-CNN: Local-aware Region CNN for Vehicle Detection in Aerial Imagery [43.91170581467171]
State-of-the-art object detection approaches have difficulties detecting dense, small targets with arbitrary orientation in large aerial images.
We present the Local-aware Region Convolutional Neural Network (LR-CNN), a novel two-stage approach for vehicle detection in aerial imagery.
arXiv Detail & Related papers (2020-05-28T19:57:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.