UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building
Instance Segmentation
- URL: http://arxiv.org/abs/2305.02627v1
- Date: Thu, 4 May 2023 08:01:38 GMT
- Title: UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building
Instance Segmentation
- Authors: Guoqing Yang, Fuyou Xue, Qi Zhang, Ke Xie, Chi-Wing Fu, Hui Huang
- Abstract summary: UrbanBIS comprises six real urban scenes, with 2.5 billion points, covering a vast area of 10.78 square kilometers.
UrbanBIS provides semantic-level annotations on a rich set of urban objects, including buildings, vehicles, vegetation, roads, and bridges.
UrbanBIS is the first 3D dataset that introduces fine-grained building sub-categories.
- Score: 50.52615875873055
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the UrbanBIS benchmark for large-scale 3D urban understanding,
supporting practical urban-level semantic and building-level instance
segmentation. UrbanBIS comprises six real urban scenes, with 2.5 billion
points, covering a vast area of 10.78 square kilometers and 3,370 buildings,
captured by 113,346 views of aerial photogrammetry. Particularly, UrbanBIS
provides not only semantic-level annotations on a rich set of urban objects,
including buildings, vehicles, vegetation, roads, and bridges, but also
instance-level annotations on the buildings. Further, UrbanBIS is the first 3D
dataset that introduces fine-grained building sub-categories, considering a
wide variety of shapes for different building types. Besides, we propose B-Seg,
a building instance segmentation method to establish UrbanBIS. B-Seg adopts an
end-to-end framework with a simple yet effective strategy for handling
large-scale point clouds. Compared with mainstream methods, B-Seg achieves
better accuracy with faster inference speed on UrbanBIS. In addition to the
carefully-annotated point clouds, UrbanBIS provides high-resolution
aerial-acquisition photos and high-quality large-scale 3D reconstruction
models, which shall facilitate a wide range of studies such as multi-view
stereo, urban LOD generation, aerial path planning, autonomous navigation, road
network extraction, and so on, thus serving as an important platform for many
intelligent city applications.
Related papers
- COHO: Context-Sensitive City-Scale Hierarchical Urban Layout Generation [1.5745692520785073]
We introduce a novel graph-based masked autoencoder (GMAE) for city-scale urban layout generation.
The method encodes attributed buildings, city blocks, communities and cities into a unified graph structure.
Our approach achieves good realism, semantic consistency, and correctness across the heterogeneous urban styles in 330 US cities.
arXiv Detail & Related papers (2024-07-16T00:49:53Z) - Urban Architect: Steerable 3D Urban Scene Generation with Layout Prior [43.14168074750301]
We introduce a compositional 3D layout representation into text-to-3D paradigm, serving as an additional prior.
It comprises a set of semantic primitives with simple geometric structures and explicit arrangement relationships.
We also present various scene editing demonstrations, showing the powers of steerable urban scene generation.
arXiv Detail & Related papers (2024-04-10T06:41:30Z) - Urban Scene Diffusion through Semantic Occupancy Map [49.20779809250597]
UrbanDiffusion is a 3D diffusion model conditioned on a Bird's-Eye View (BEV) map.
Our model learns the data distribution of scene-level structures within a latent space.
After training on real-world driving datasets, our model can generate a wide range of diverse urban scenes.
arXiv Detail & Related papers (2024-03-18T11:54:35Z) - UrbanKGent: A Unified Large Language Model Agent Framework for Urban Knowledge Graph Construction [4.829677240798159]
Urban knowledge graph construction (UrbanKGC) still heavily relies on manual effort, hindering its potential advancement.
This paper presents UrbanKGent, a unified large language model agent framework, for urban knowledge graph construction.
arXiv Detail & Related papers (2024-02-10T01:50:19Z) - Sat2Scene: 3D Urban Scene Generation from Satellite Images with Diffusion [77.34078223594686]
We propose a novel architecture for direct 3D scene generation by introducing diffusion models into 3D sparse representations and combining them with neural rendering techniques.
Specifically, our approach generates texture colors at the point level for a given geometry using a 3D diffusion model first, which is then transformed into a scene representation in a feed-forward manner.
Experiments in two city-scale datasets show that our model demonstrates proficiency in generating photo-realistic street-view image sequences and cross-view urban scenes from satellite imagery.
arXiv Detail & Related papers (2024-01-19T16:15:37Z) - Building3D: An Urban-Scale Dataset and Benchmarks for Learning Roof
Structures from Point Clouds [4.38301148531795]
Existing datasets for 3D modeling mainly focus on common objects such as furniture or cars.
We present a urban-scale dataset consisting of more than 160 thousands buildings along with corresponding point clouds, mesh and wire-frame models, covering 16 cities in Estonia about 998 Km2.
Experimental results indicate that Building3D has challenges of high intra-class variance, data imbalance and large-scale noises.
arXiv Detail & Related papers (2023-07-21T21:38:57Z) - SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point
Clouds [52.624157840253204]
We introduce SensatUrban, an urban-scale UAV photogrammetry point cloud dataset consisting of nearly three billion points collected from three UK cities, covering 7.6 km2.
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
arXiv Detail & Related papers (2022-01-12T14:48:11Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset,
Benchmarks and Challenges [52.624157840253204]
We present an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points.
Our dataset consists of large areas from three UK cities, covering about 7.6 km2 of the city landscape.
We evaluate the performance of state-of-the-art algorithms on our dataset and provide a comprehensive analysis of the results.
arXiv Detail & Related papers (2020-09-07T14:47:07Z) - Deep Learning Guided Building Reconstruction from Satellite
Imagery-derived Point Clouds [39.36437891978871]
We present a reliable and effective approach for building model reconstruction from the point clouds generated from satellite images.
Specifically, a deep-learning approach is adopted to distinguish the shape of building roofs in complex and yet noisy scenes.
As the first effort to address the public need of large scale city model generation, the development is deployed as open source software.
arXiv Detail & Related papers (2020-05-19T05:38:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.