HRHD-HK: A benchmark dataset of high-rise and high-density urban scenes
for 3D semantic segmentation of photogrammetric point clouds
- URL: http://arxiv.org/abs/2307.07976v2
- Date: Mon, 11 Dec 2023 03:28:47 GMT
- Title: HRHD-HK: A benchmark dataset of high-rise and high-density urban scenes
for 3D semantic segmentation of photogrammetric point clouds
- Authors: Maosu Li, Yijie Wu, Anthony G.O. Yeh, Fan Xue
- Abstract summary: This paper presents a benchmark dataset of high-rise urban point clouds, namely High-Rise, High-Density urban scenes of Hong Kong (HRHD-HK)
HRHD-HK arranged in 150 tiles contains 273 million colorful photogrammetric 3D points from diverse urban settings.
This paper also comprehensively evaluates eight popular semantic segmentation methods on the HRHD-HK dataset.
- Score: 0.40964539027092917
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Many existing 3D semantic segmentation methods, deep learning in computer
vision notably, claimed to achieve desired results on urban point clouds. Thus,
it is significant to assess these methods quantitatively in diversified
real-world urban scenes, encompassing high-rise, low-rise, high-density, and
low-density urban areas. However, existing public benchmark datasets primarily
represent low-rise scenes from European cities and cannot assess the methods
comprehensively. This paper presents a benchmark dataset of high-rise urban
point clouds, namely High-Rise, High-Density urban scenes of Hong Kong
(HRHD-HK). HRHD-HK arranged in 150 tiles contains 273 million colorful
photogrammetric 3D points from diverse urban settings. The semantic labels of
HRHD-HK include building, vegetation, road, waterbody, facility, terrain, and
vehicle. To our best knowledge, HRHD-HK is the first photogrammetric dataset
that focuses on HRHD urban areas. This paper also comprehensively evaluates
eight popular semantic segmentation methods on the HRHD-HK dataset.
Experimental results confirmed plenty of room for enhancing the current 3D
semantic segmentation of point clouds, especially for city objects with small
volumes. Our dataset is publicly available at
https://doi.org/10.25442/hku.23701866.v2.
Related papers
- 3D Question Answering for City Scene Understanding [12.433903847890322]
3D multimodal question answering (MQA) plays a crucial role in scene understanding by enabling intelligent agents to comprehend their surroundings in 3D environments.
We introduce a novel 3D MQA dataset named City-3DQA for city-level scene understanding.
A new benchmark is reported and our proposed Sg-CityU achieves accuracy of 63.94 % and 63.76 % in different settings of City-3DQA.
arXiv Detail & Related papers (2024-07-24T16:22:27Z) - HUGS: Holistic Urban 3D Scene Understanding via Gaussian Splatting [53.6394928681237]
holistic understanding of urban scenes based on RGB images is a challenging yet important problem.
Our main idea involves the joint optimization of geometry, appearance, semantics, and motion using a combination of static and dynamic 3D Gaussians.
Our approach offers the ability to render new viewpoints in real-time, yielding 2D and 3D semantic information with high accuracy.
arXiv Detail & Related papers (2024-03-19T13:39:05Z) - CityRefer: Geography-aware 3D Visual Grounding Dataset on City-scale
Point Cloud Data [15.526523262690965]
We introduce the CityRefer dataset for city-level visual grounding.
The dataset consists of 35k natural language descriptions of 3D objects appearing in SensatUrban city scenes and 5k landmarks labels synchronizing with OpenStreetMap.
arXiv Detail & Related papers (2023-10-28T18:05:32Z) - Cross-City Matters: A Multimodal Remote Sensing Benchmark Dataset for
Cross-City Semantic Segmentation using High-Resolution Domain Adaptation
Networks [82.82866901799565]
We build a new set of multimodal remote sensing benchmark datasets (including hyperspectral, multispectral, SAR) for the study purpose of the cross-city semantic segmentation task.
Beyond the single city, we propose a high-resolution domain adaptation network, HighDAN, to promote the AI model's generalization ability from the multi-city environments.
HighDAN is capable of retaining the spatially topological structure of the studied urban scene well in a parallel high-to-low resolution fusion fashion.
arXiv Detail & Related papers (2023-09-26T23:55:39Z) - SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point
Clouds [52.624157840253204]
We introduce SensatUrban, an urban-scale UAV photogrammetry point cloud dataset consisting of nearly three billion points collected from three UK cities, covering 7.6 km2.
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
arXiv Detail & Related papers (2022-01-12T14:48:11Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset,
Benchmarks and Challenges [52.624157840253204]
We present an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points.
Our dataset consists of large areas from three UK cities, covering about 7.6 km2 of the city landscape.
We evaluate the performance of state-of-the-art algorithms on our dataset and provide a comprehensive analysis of the results.
arXiv Detail & Related papers (2020-09-07T14:47:07Z) - HDNet: Human Depth Estimation for Multi-Person Camera-Space Localization [83.57863764231655]
We propose the Human Depth Estimation Network (HDNet), an end-to-end framework for absolute root joint localization.
A skeleton-based Graph Neural Network (GNN) is utilized to propagate features among joints.
We evaluate our HDNet on the root joint localization and root-relative 3D pose estimation tasks with two benchmark datasets.
arXiv Detail & Related papers (2020-07-17T12:44:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.