FusionLane: Multi-Sensor Fusion for Lane Marking Semantic Segmentation
Using Deep Neural Networks
- URL: http://arxiv.org/abs/2003.04404v1
- Date: Mon, 9 Mar 2020 20:33:30 GMT
- Title: FusionLane: Multi-Sensor Fusion for Lane Marking Semantic Segmentation
Using Deep Neural Networks
- Authors: Ruochen Yin, Biao Yu, Huapeng Wu, Yutao Song, Runxin Niu
- Abstract summary: This paper proposes a lane marking semantic segmentation method based on LIDAR and camera fusion deep neural network.
Experiments on more than 14,000 image datasets have shown the proposed method has better performance on the semantic segmentation of the points cloud bird's eye view.
- Score: 1.0062127381149395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is a crucial step to achieve effective semantic segmentation of lane
marking during the construction of the lane level high-precision map. In recent
years, many image semantic segmentation methods have been proposed. These
methods mainly focus on the image from camera, due to the limitation of the
sensor itself, the accurate three-dimensional spatial position of the lane
marking cannot be obtained, so the demand for the lane level high-precision map
construction cannot be met. This paper proposes a lane marking semantic
segmentation method based on LIDAR and camera fusion deep neural network.
Different from other methods, in order to obtain accurate position information
of the segmentation results, the semantic segmentation object of this paper is
a bird's eye view converted from a LIDAR points cloud instead of an image
captured by a camera. This method first uses the deeplabv3+ [\ref{ref:1}]
network to segment the image captured by the camera, and the segmentation
result is merged with the point clouds collected by the LIDAR as the input of
the proposed network. In this neural network, we also add a long short-term
memory (LSTM) structure to assist the network for semantic segmentation of lane
markings by using the the time series information. The experiments on more than
14,000 image datasets which we have manually labeled and expanded have shown
the proposed method has better performance on the semantic segmentation of the
points cloud bird's eye view. Therefore, the automation of high-precision map
construction can be significantly improved. Our code is available at
https://github.com/rolandying/FusionLane.
Related papers
- Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - De-coupling and De-positioning Dense Self-supervised Learning [65.56679416475943]
Dense Self-Supervised Learning (SSL) methods address the limitations of using image-level feature representations when handling images with multiple objects.
We show that they suffer from coupling and positional bias, which arise from the receptive field increasing with layer depth and zero-padding.
We demonstrate the benefits of our method on COCO and on a new challenging benchmark, OpenImage-MINI, for object classification, semantic segmentation, and object detection.
arXiv Detail & Related papers (2023-03-29T18:07:25Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - EAA-Net: Rethinking the Autoencoder Architecture with Intra-class
Features for Medical Image Segmentation [4.777011444412729]
We propose a light-weight end-to-end segmentation framework based on multi-task learning, termed Edge Attention autoencoder Network (EAA-Net)
Our approach not only utilizes the segmentation network to obtain inter-class features, but also applies the reconstruction network to extract intra-class features among the foregrounds.
Experimental results show that our method performs well in medical image segmentation tasks.
arXiv Detail & Related papers (2022-08-19T07:42:55Z) - From Explanations to Segmentation: Using Explainable AI for Image
Segmentation [1.8581514902689347]
We build upon the advances of the Explainable AI (XAI) community and extract a pixel-wise binary segmentation.
We show that we achieve similar results compared to an established U-Net segmentation architecture.
The proposed method can be trained in a weakly supervised fashion, as the training samples must be only labeled on image-level.
arXiv Detail & Related papers (2022-02-01T10:26:10Z) - Multi-Scale Feature Fusion: Learning Better Semantic Segmentation for
Road Pothole Detection [9.356003255288417]
This paper presents a novel pothole detection approach based on single-modal semantic segmentation.
It first extracts visual features from input images using a convolutional neural network.
A channel attention module then reweighs the channel features to enhance the consistency of different feature maps.
arXiv Detail & Related papers (2021-12-24T15:07:47Z) - A Novel Upsampling and Context Convolution for Image Semantic
Segmentation [0.966840768820136]
Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
arXiv Detail & Related papers (2021-03-20T06:16:42Z) - Semantic Segmentation With Multi Scale Spatial Attention For Self
Driving Cars [2.7317088388886384]
We present a novel neural network using multi scale feature fusion at various scales for accurate and efficient semantic image segmentation.
We used ResNet based feature extractor, dilated convolutional layers in downsampling part, atrous convolutional layers in the upsampling part and used concat operation to merge them.
A new attention module is proposed to encode more contextual information and enhance the receptive field of the network.
arXiv Detail & Related papers (2020-06-30T20:19:09Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z) - Real-Time High-Performance Semantic Image Segmentation of Urban Street
Scenes [98.65457534223539]
We propose a real-time high-performance DCNN-based method for robust semantic segmentation of urban street scenes.
The proposed method achieves the accuracy of 73.6% and 68.0% mean Intersection over Union (mIoU) with the inference speed of 51.0 fps and 39.3 fps.
arXiv Detail & Related papers (2020-03-11T08:45:53Z) - RGB-based Semantic Segmentation Using Self-Supervised Depth Pre-Training [77.62171090230986]
We propose an easily scalable and self-supervised technique that can be used to pre-train any semantic RGB segmentation method.
In particular, our pre-training approach makes use of automatically generated labels that can be obtained using depth sensors.
We show how our proposed self-supervised pre-training with HN-labels can be used to replace ImageNet pre-training.
arXiv Detail & Related papers (2020-02-06T11:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.