CG-Net: Conditional GIS-aware Network for Individual Building
Segmentation in VHR SAR Images
- URL: http://arxiv.org/abs/2011.08362v1
- Date: Tue, 17 Nov 2020 01:52:22 GMT
- Title: CG-Net: Conditional GIS-aware Network for Individual Building
Segmentation in VHR SAR Images
- Authors: Yao Sun, Yuansheng Hua, Lichao Mou, Xiao Xiang Zhu
- Abstract summary: This paper addresses the issue of individual building segmentation from a single VHR SAR image in large-scale urban areas.
We introduce building footprints from GIS data as complementary information and propose a novel conditional GIS-aware network (CG-Net)
The proposed model learns multi-level visual features and employs building footprints to normalize the features for predicting building masks in the SAR image.
- Score: 25.87229252642239
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object retrieval and reconstruction from very high resolution (VHR) synthetic
aperture radar (SAR) images are of great importance for urban SAR applications,
yet highly challenging owing to the complexity of SAR data. This paper
addresses the issue of individual building segmentation from a single VHR SAR
image in large-scale urban areas. To achieve this, we introduce building
footprints from GIS data as complementary information and propose a novel
conditional GIS-aware network (CG-Net). The proposed model learns multi-level
visual features and employs building footprints to normalize the features for
predicting building masks in the SAR image. We validate our method using a high
resolution spotlight TerraSAR-X image collected over Berlin. Experimental
results show that the proposed CG-Net effectively brings improvements with
variant backbones. We further compare two representations of building
footprints, namely complete building footprints and sensor-visible footprint
segments, for our task, and conclude that the use of the former leads to better
segmentation results. Moreover, we investigate the impact of inaccurate GIS
data on our CG-Net, and this study shows that CG-Net is robust against
positioning errors in GIS data. In addition, we propose an approach of ground
truth generation of buildings from an accurate digital elevation model (DEM),
which can be used to generate large-scale SAR image datasets. The segmentation
results can be applied to reconstruct 3D building models at level-of-detail
(LoD) 1, which is demonstrated in our experiments.
Related papers
- Rethinking Image Super-Resolution from Training Data Perspectives [54.28824316574355]
We investigate the understudied effect of the training data used for image super-resolution (SR)
With this, we propose an automated image evaluation pipeline.
We find that datasets with (i) low compression artifacts, (ii) high within-image diversity as judged by the number of different objects, and (iii) a large number of images from ImageNet or PASS all positively affect SR performance.
arXiv Detail & Related papers (2024-09-01T16:25:04Z) - GeoLRM: Geometry-Aware Large Reconstruction Model for High-Quality 3D Gaussian Generation [65.33726478659304]
We introduce the Geometry-Aware Large Reconstruction Model (GeoLRM), an approach which can predict high-quality assets with 512k Gaussians and 21 input images in only 11 GB GPU memory.
Previous works neglect the inherent sparsity of 3D structure and do not utilize explicit geometric relationships between 3D and 2D images.
GeoLRM tackles these issues by incorporating a novel 3D-aware transformer structure that directly processes 3D points and uses deformable cross-attention mechanisms.
arXiv Detail & Related papers (2024-06-21T17:49:31Z) - S2RC-GCN: A Spatial-Spectral Reliable Contrastive Graph Convolutional Network for Complex Land Cover Classification Using Hyperspectral Images [10.579474650543471]
This study proposes a novel spatial-spectral reliable contrastive graph convolutional classification framework named S2RC-GCN.
Specifically, we fused the spectral and spatial features extracted by the 1D- and 2D-encoder, and the 2D-encoder includes an attention model to automatically extract important information.
We then leveraged the fused high-level features to construct graphs and fed the resulting graphs into the GCNs to determine more effective graph representations.
arXiv Detail & Related papers (2024-04-01T07:17:02Z) - SARDet-100K: Towards Open-Source Benchmark and ToolKit for Large-Scale SAR Object Detection [79.23689506129733]
We establish a new benchmark dataset and an open-source method for large-scale SAR object detection.
Our dataset, SARDet-100K, is a result of intense surveying, collecting, and standardizing 10 existing SAR detection datasets.
To the best of our knowledge, SARDet-100K is the first COCO-level large-scale multi-class SAR object detection dataset ever created.
arXiv Detail & Related papers (2024-03-11T09:20:40Z) - Rotated Multi-Scale Interaction Network for Referring Remote Sensing Image Segmentation [63.15257949821558]
Referring Remote Sensing Image (RRSIS) is a new challenge that combines computer vision and natural language processing.
Traditional Referring Image (RIS) approaches have been impeded by the complex spatial scales and orientations found in aerial imagery.
We introduce the Rotated Multi-Scale Interaction Network (RMSIN), an innovative approach designed for the unique demands of RRSIS.
arXiv Detail & Related papers (2023-12-19T08:14:14Z) - Building Extraction from Remote Sensing Images via an Uncertainty-Aware
Network [18.365220543556113]
Building extraction plays an essential role in many applications, such as city planning and urban dynamic monitoring.
We propose a novel and straightforward Uncertainty-Aware Network (UANet) to alleviate this problem.
Results demonstrate that the proposed UANet outperforms other state-of-the-art algorithms by a large margin.
arXiv Detail & Related papers (2023-07-23T12:42:15Z) - Highly Accurate Dichotomous Image Segmentation [139.79513044546]
A new task called dichotomous image segmentation (DIS) aims to segment highly accurate objects from natural images.
We collect the first large-scale dataset, DIS5K, which contains 5,470 high-resolution (e.g., 2K, 4K or larger) images.
We also introduce a simple intermediate supervision baseline (IS-Net) using both feature-level and mask-level guidance for DIS model training.
arXiv Detail & Related papers (2022-03-06T20:09:19Z) - Large-scale Building Height Retrieval from Single SAR Imagery based on
Bounding Box Regression Networks [21.788338971571736]
Building height retrieval from synthetic aperture radar (SAR) imagery is of great importance for urban applications.
This paper addresses the issue of building height retrieval in large-scale urban areas from a single TerraSAR-X spotlight or stripmap image.
arXiv Detail & Related papers (2021-11-18T00:39:48Z) - Sci-Net: a Scale Invariant Model for Building Detection from Aerial
Images [0.0]
We propose a Scale-invariant neural network (Sci-Net) that is able to segment buildings present in aerial images at different spatial resolutions.
Specifically, we modified the U-Net architecture and fused it with dense Atrous Spatial Pyramid Pooling (ASPP) to extract fine-grained multi-scale representations.
arXiv Detail & Related papers (2021-11-12T16:45:20Z) - Ground material classification and for UAV-based photogrammetric 3D data
A 2D-3D Hybrid Approach [1.3359609092684614]
In recent years, photogrammetry has been widely used in many areas to create 3D virtual data representing the physical environment.
These cutting-edge technologies have caught the US Army and Navy's attention for the purpose of rapid 3D battlefield reconstruction, virtual training, and simulations.
arXiv Detail & Related papers (2021-09-24T22:29:26Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.