LGENet: Local and Global Encoder Network for Semantic Segmentation of
Airborne Laser Scanning Point Clouds
- URL: http://arxiv.org/abs/2012.10192v1
- Date: Fri, 18 Dec 2020 12:26:53 GMT
- Title: LGENet: Local and Global Encoder Network for Semantic Segmentation of
Airborne Laser Scanning Point Clouds
- Authors: Yaping Lin, George Vosselman, Yanpeng Cao, Michael Ying Yang
- Abstract summary: We present a local and global encoder network (LGENet) for semantic segmentation of ALS point clouds.
For the ISPRS benchmark dataset, our model achieves state-of-the-art results with an overall accuracy of 0.845 and an average F1 score of 0.737.
- Score: 17.840158282335874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interpretation of Airborne Laser Scanning (ALS) point clouds is a critical
procedure for producing various geo-information products like 3D city models,
digital terrain models and land use maps. In this paper, we present a local and
global encoder network (LGENet) for semantic segmentation of ALS point clouds.
Adapting the KPConv network, we first extract features by both 2D and 3D point
convolutions to allow the network to learn more representative local geometry.
Then global encoders are used in the network to exploit contextual information
at the object and point level. We design a segment-based Edge Conditioned
Convolution to encode the global context between segments. We apply a
spatial-channel attention module at the end of the network, which not only
captures the global interdependencies between points but also models
interactions between channels. We evaluate our method on two ALS datasets
namely, the ISPRS benchmark dataset and DCF2019 dataset. For the ISPRS
benchmark dataset, our model achieves state-of-the-art results with an overall
accuracy of 0.845 and an average F1 score of 0.737. With regards to the DFC2019
dataset, our proposed network achieves an overall accuracy of 0.984 and an
average F1 score of 0.834.
Related papers
- Low-Resource White-Box Semantic Segmentation of Supporting Towers on 3D
Point Clouds via Signature Shape Identification [52.77024349608834]
SCENE-Net is a low-resource white-box model for 3D point cloud semantic segmentation.
Our training time on a laptop is 85min, and our inference time is 20ms.
We release a 40000 Km labeled dataset of rural terrain point clouds and our code implementation.
arXiv Detail & Related papers (2023-06-13T14:36:06Z) - Dynamic Clustering Transformer Network for Point Cloud Segmentation [23.149220817575195]
We propose a novel 3D point cloud representation network, called Dynamic Clustering Transformer Network (DCTNet)
It has an encoder-decoder architecture, allowing for both local and global feature learning.
Our method was evaluated on an object-based dataset (ShapeNet), an urban navigation dataset (Toronto-3D), and a multispectral LiDAR dataset.
arXiv Detail & Related papers (2023-05-30T01:11:05Z) - LACV-Net: Semantic Segmentation of Large-Scale Point Cloud Scene via
Local Adaptive and Comprehensive VLAD [13.907586081922345]
We propose an end-to-end deep neural network called LACV-Net for large-scale point cloud semantic segmentation.
The proposed network contains three main components: 1) a local adaptive feature augmentation module (LAFA) to adaptively learn the similarity of centroids and neighboring points to augment the local context; 2) a comprehensive VLAD module that fuses local features with multi-layer, multi-scale, and multi-resolution to represent a comprehensive global description vector; and 3) an aggregation loss function to effectively optimize the segmentation boundaries by constraining the adaptive weight from the LAFA module.
arXiv Detail & Related papers (2022-10-12T02:11:00Z) - Global Hierarchical Attention for 3D Point Cloud Analysis [88.56041763189162]
We propose a new attention mechanism, called Global Hierarchical Attention (GHA) for 3D point cloud analysis.
For the task of semantic segmentation, GHA gives a +1.7% mIoU increase to the MinkowskiEngine baseline on ScanNet.
For the 3D object detection task, GHA improves the CenterPoint baseline by +0.5% mAP on the nuScenes dataset.
arXiv Detail & Related papers (2022-08-07T19:16:30Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - Neural RF SLAM for unsupervised positioning and mapping with channel
state information [51.484516640867525]
We present a neural network architecture for jointly learning user locations and environment mapping up to isometry.
The proposed model learns an interpretable latent, i.e., user location, by just enforcing a physics-based decoder.
arXiv Detail & Related papers (2022-03-15T21:32:44Z) - TC-Net: Triple Context Network for Automated Stroke Lesion Segmentation [0.5482532589225552]
We propose a new network, Triple Context Network (TC-Net), with the capture of spatial contextual information as the core.
Our network is evaluated on the open dataset ATLAS, achieving the highest score of 0.594, Hausdorff distance of 27.005 mm, and average symmetry surface distance of 7.137 mm.
arXiv Detail & Related papers (2022-02-28T11:12:16Z) - PRA-Net: Point Relation-Aware Network for 3D Point Cloud Analysis [56.91758845045371]
We propose a novel framework named Point Relation-Aware Network (PRA-Net)
It is composed of an Intra-region Structure Learning (ISL) module and an Inter-region Relation Learning (IRL) module.
Experiments on several 3D benchmarks covering shape classification, keypoint estimation, and part segmentation have verified the effectiveness and the ability of PRA-Net.
arXiv Detail & Related papers (2021-12-09T13:24:43Z) - An Attention-Fused Network for Semantic Segmentation of
Very-High-Resolution Remote Sensing Imagery [26.362854938949923]
We propose a novel convolutional neural network architecture, named attention-fused network (AFNet)
We achieve state-of-the-art performance with an overall accuracy of 91.7% and a mean F1 score of 90.96% on the ISPRS Vaihingen 2D dataset and the ISPRS Potsdam 2D dataset.
arXiv Detail & Related papers (2021-05-10T06:23:27Z) - PIG-Net: Inception based Deep Learning Architecture for 3D Point Cloud
Segmentation [0.9137554315375922]
We propose a inception based deep network architecture called PIG-Net, that effectively characterizes the local and global geometric details of the point clouds.
We perform an exhaustive experimental analysis of the PIG-Net architecture on two state-of-the-art datasets.
arXiv Detail & Related papers (2021-01-28T13:27:55Z) - Interactive Video Object Segmentation Using Global and Local Transfer
Modules [51.93009196085043]
We develop a deep neural network, which consists of the annotation network (A-Net) and the transfer network (T-Net)
Given user scribbles on a frame, A-Net yields a segmentation result based on the encoder-decoder architecture.
We train the entire network in two stages, by emulating user scribbles and employing an auxiliary loss.
arXiv Detail & Related papers (2020-07-16T06:49:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.