Adversarial Shape Learning for Building Extraction in VHR Remote Sensing
Images
- URL: http://arxiv.org/abs/2102.11262v2
- Date: Thu, 25 Feb 2021 13:58:51 GMT
- Title: Adversarial Shape Learning for Building Extraction in VHR Remote Sensing
Images
- Authors: Lei Ding, Hao Tang, Yahui Liu, Yilei Shi and Lorenzo Bruzzone
- Abstract summary: We propose an adversarial shape learning network (ASLNet) to model the building shape patterns.
Experiments show that the proposed ASLNet improves both the pixel-based accuracy and the object-based measurements by a large margin.
- Score: 18.650642666164252
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building extraction in VHR RSIs remains to be a challenging task due to
occlusion and boundary ambiguity problems. Although conventional convolutional
neural networks (CNNs) based methods are capable of exploiting local texture
and context information, they fail to capture the shape patterns of buildings,
which is a necessary constraint in the human recognition. In this context, we
propose an adversarial shape learning network (ASLNet) to model the building
shape patterns, thus improving the accuracy of building segmentation. In the
proposed ASLNet, we introduce the adversarial learning strategy to explicitly
model the shape constraints, as well as a CNN shape regularizer to strengthen
the embedding of shape features. To assess the geometric accuracy of building
segmentation results, we further introduced several object-based assessment
metrics. Experiments on two open benchmark datasets show that the proposed
ASLNet improves both the pixel-based accuracy and the object-based measurements
by a large margin. The code is available at: https://github.com/ggsDing/ASLNet
Related papers
- Segmenting objects with Bayesian fusion of active contour models and convnet priors [0.729597981661727]
We propose a novel instance segmentation method geared towards Natural Resource Monitoring (NRM) imagery.
We formulate the problem as Bayesian maximum a posteriori inference which, in learning the individual object contours, incorporates shape, location, and position priors.
In experiments, we tackle the challenging, real-world problem of segmenting individual dead tree crowns and precise contours.
arXiv Detail & Related papers (2024-10-09T20:36:43Z) - Unsupervised Non-Rigid Point Cloud Matching through Large Vision Models [1.3030624795284795]
We propose a learning-based framework for non-rigid point cloud matching.
Key insight is to incorporate semantic features derived from large vision models (LVMs)
Our framework effectively leverages the structural information contained in the semantic features to address ambiguities arise from self-similarities among local geometries.
arXiv Detail & Related papers (2024-08-16T07:02:19Z) - Shape Anchor Guided Holistic Indoor Scene Understanding [9.463220988312218]
We propose a shape anchor guided learning strategy (AncLearn) for robust holistic indoor scene understanding.
AncLearn generates anchors that dynamically fit instance surfaces to (i) unmix noise and target-related features for offering reliable proposals at the detection stage.
We embed AncLearn into a reconstruction-from-detection learning system (AncRec) to generate high-quality semantic scene models.
arXiv Detail & Related papers (2023-09-20T08:30:20Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - A Convolutional Neural Network Approach to the Classification of
Engineering Models [0.9558392439655015]
This paper presents a deep learning approach for the classification of Engineering (CAD) models using Convolutional Neural Networks (CNNs)
It is proposed to use a residual network architecture for CADNET, inspired by the popular ResNet.
The LFD-based CNN approach using the proposed network architecture, along with gradient boosting yielded the best classification accuracy on CADNET.
arXiv Detail & Related papers (2021-07-14T04:33:50Z) - Learning Geometry-Disentangled Representation for Complementary
Understanding of 3D Object Point Cloud [50.56461318879761]
We propose Geometry-Disentangled Attention Network (GDANet) for 3D image processing.
GDANet disentangles point clouds into contour and flat part of 3D objects, respectively denoted by sharp and gentle variation components.
Experiments on 3D object classification and segmentation benchmarks demonstrate that GDANet achieves the state-of-the-arts with fewer parameters.
arXiv Detail & Related papers (2020-12-20T13:35:00Z) - Local Grid Rendering Networks for 3D Object Detection in Point Clouds [98.02655863113154]
CNNs are powerful but it would be computationally costly to directly apply convolutions on point data after voxelizing the entire point clouds to a dense regular 3D grid.
We propose a novel and principled Local Grid Rendering (LGR) operation to render the small neighborhood of a subset of input points into a low-resolution 3D grid independently.
We validate LGR-Net for 3D object detection on the challenging ScanNet and SUN RGB-D datasets.
arXiv Detail & Related papers (2020-07-04T13:57:43Z) - Shape-Oriented Convolution Neural Network for Point Cloud Analysis [59.405388577930616]
Point cloud is a principal data structure adopted for 3D geometric information encoding.
Shape-oriented message passing scheme dubbed ShapeConv is proposed to focus on the representation learning of the underlying shape formed by each local neighboring point.
arXiv Detail & Related papers (2020-04-20T16:11:51Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.