Deep Learning Guided Building Reconstruction from Satellite
Imagery-derived Point Clouds
- URL: http://arxiv.org/abs/2005.09223v1
- Date: Tue, 19 May 2020 05:38:06 GMT
- Title: Deep Learning Guided Building Reconstruction from Satellite
Imagery-derived Point Clouds
- Authors: Bo Xu, Xu Zhang, Zhixin Li, Matt Leotta, Shih-Fu Chang, Jie Shan
- Abstract summary: We present a reliable and effective approach for building model reconstruction from the point clouds generated from satellite images.
Specifically, a deep-learning approach is adopted to distinguish the shape of building roofs in complex and yet noisy scenes.
As the first effort to address the public need of large scale city model generation, the development is deployed as open source software.
- Score: 39.36437891978871
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D urban reconstruction of buildings from remotely sensed imagery has drawn
significant attention during the past two decades. While aerial imagery and
LiDAR provide higher resolution, satellite imagery is cheaper and more
efficient to acquire for large scale need. However, the high, orbital altitude
of satellite observation brings intrinsic challenges, like unpredictable
atmospheric effect, multi view angles, significant radiometric differences due
to the necessary multiple views, diverse land covers and urban structures in a
scene, small base-height ratio or narrow field of view, all of which may
degrade 3D reconstruction quality. To address these major challenges, we
present a reliable and effective approach for building model reconstruction
from the point clouds generated from multi-view satellite images. We utilize
multiple types of primitive shapes to fit the input point cloud. Specifically,
a deep-learning approach is adopted to distinguish the shape of building roofs
in complex and yet noisy scenes. For points that belong to the same roof shape,
a multi-cue, hierarchical RANSAC approach is proposed for efficient and
reliable segmenting and reconstructing the building point cloud. Experimental
results over four selected urban areas (0.34 to 2.04 sq km in size) demonstrate
the proposed method can generate detailed roof structures under noisy data
environments. The average successful rate for building shape recognition is
83.0%, while the overall completeness and correctness are over 70% with
reference to ground truth created from airborne lidar. As the first effort to
address the public need of large scale city model generation, the development
is deployed as open source software.
Related papers
- APC2Mesh: Bridging the gap from occluded building façades to full 3D models [5.113739955215433]
We propose APC2Mesh which integrates point completion into a 3D reconstruction pipeline.
Specifically, we leveraged complete points generated from occluded ones as input to a linearized skip attention-based deformation network for 3D mesh reconstruction.
arXiv Detail & Related papers (2024-04-03T01:29:30Z) - Aerial Lifting: Neural Urban Semantic and Building Instance Lifting from Aerial Imagery [51.73680703579997]
We present a neural radiance field method for urban-scale semantic and building-level instance segmentation from aerial images.
objects in urban aerial images exhibit substantial variations in size, including buildings, cars, and roads.
We introduce a scale-adaptive semantic label fusion strategy that enhances the segmentation of objects of varying sizes.
We then introduce a novel cross-view instance label grouping strategy to mitigate the multi-view inconsistency problem in the 2D instance labels.
arXiv Detail & Related papers (2024-03-18T14:15:39Z) - Point2Building: Reconstructing Buildings from Airborne LiDAR Point Clouds [23.897507889025817]
We present a learning-based approach to reconstruct buildings as 3D polygonal meshes from airborne LiDAR point clouds.
Our model learns directly from the point cloud data, thereby reducing error propagation and increasing the fidelity of the reconstruction.
We experimentally validate our method on a collection of airborne LiDAR data of Zurich, Berlin and Tallinn.
arXiv Detail & Related papers (2024-03-04T15:46:50Z) - Sat2Scene: 3D Urban Scene Generation from Satellite Images with Diffusion [77.34078223594686]
We propose a novel architecture for direct 3D scene generation by introducing diffusion models into 3D sparse representations and combining them with neural rendering techniques.
Specifically, our approach generates texture colors at the point level for a given geometry using a 3D diffusion model first, which is then transformed into a scene representation in a feed-forward manner.
Experiments in two city-scale datasets show that our model demonstrates proficiency in generating photo-realistic street-view image sequences and cross-view urban scenes from satellite imagery.
arXiv Detail & Related papers (2024-01-19T16:15:37Z) - Patch-Wise Point Cloud Generation: A Divide-and-Conquer Approach [83.05340155068721]
We devise a new 3d point cloud generation framework using a divide-and-conquer approach.
All patch generators are based on learnable priors, which aim to capture the information of geometry primitives.
Experimental results on a variety of object categories from the most popular point cloud dataset, ShapeNet, show the effectiveness of the proposed patch-wise point cloud generation.
arXiv Detail & Related papers (2023-07-22T11:10:39Z) - UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building
Instance Segmentation [50.52615875873055]
UrbanBIS comprises six real urban scenes, with 2.5 billion points, covering a vast area of 10.78 square kilometers.
UrbanBIS provides semantic-level annotations on a rich set of urban objects, including buildings, vehicles, vegetation, roads, and bridges.
UrbanBIS is the first 3D dataset that introduces fine-grained building sub-categories.
arXiv Detail & Related papers (2023-05-04T08:01:38Z) - Building Coverage Estimation with Low-resolution Remote Sensing Imagery [65.95520230761544]
We propose a method for estimating building coverage using only publicly available low-resolution satellite imagery.
Our model achieves a coefficient of determination as high as 0.968 on predicting building coverage in regions of different levels of development around the world.
arXiv Detail & Related papers (2023-01-04T05:19:33Z) - sat2pc: Estimating Point Cloud of Building Roofs from 2D Satellite
Images [1.8884278918443564]
We propose sat2pc, a deep learning architecture that predicts the point of a building roof from a single 2D satellite image.
Our results show that sat2pc was able to outperform existing baselines by at least 18.6%.
arXiv Detail & Related papers (2022-05-25T03:24:40Z) - City3D: Large-scale Urban Reconstruction from Airborne Point Clouds [3.7422375336465037]
We present a fully automatic approach for reconstructing compact 3D building models from large-scale airborne point clouds.
Based on the observation that urban buildings typically consist of planar roofs connected with vertical walls to the ground, we propose an approach to infer the vertical walls directly from the data.
Experiments on various large-scale airborne point clouds have demonstrated that the method is superior to the state-of-the-art methods in terms of reconstruction accuracy and robustness.
arXiv Detail & Related papers (2022-01-25T12:41:11Z) - Boundary Regularized Building Footprint Extraction From Satellite Images
Using Deep Neural Network [6.371173732947292]
We propose a novel deep neural network, which enables to jointly detect building instance and regularize noisy building boundary shapes from a single satellite imagery.
Our model can accomplish multi-tasks of object localization, recognition, semantic labelling and geometric shape extraction simultaneously.
arXiv Detail & Related papers (2020-06-23T17:24:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.