HoliCity: A City-Scale Data Platform for Learning Holistic 3D Structures
- URL: http://arxiv.org/abs/2008.03286v2
- Date: Thu, 25 Mar 2021 05:06:19 GMT
- Title: HoliCity: A City-Scale Data Platform for Learning Holistic 3D Structures
- Authors: Yichao Zhou, Jingwei Huang, Xili Dai, Shichen Liu, Linjie Luo, Zhili
Chen, Yi Ma
- Abstract summary: This dataset has 6,300 real-world panoramas that are accurately aligned with a CAD model downtown London with an area of 20 km2 times.
The ultimate goal of this dataset is to support real applications for city reconstruction, mapping, and augmented reality.
- Score: 39.2984574045825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present HoliCity, a city-scale 3D dataset with rich structural
information. Currently, this dataset has 6,300 real-world panoramas of
resolution $13312 \times 6656$ that are accurately aligned with the CAD model
of downtown London with an area of more than 20 km$^2$, in which the median
reprojection error of the alignment of an average image is less than half a
degree. This dataset aims to be an all-in-one data platform for research of
learning abstracted high-level holistic 3D structures that can be derived from
city CAD models, e.g., corners, lines, wireframes, planes, and cuboids, with
the ultimate goal of supporting real-world applications including city-scale
reconstruction, localization, mapping, and augmented reality. The accurate
alignment of the 3D CAD models and panoramas also benefits low-level 3D vision
tasks such as surface normal estimation, as the surface normal extracted from
previous LiDAR-based datasets is often noisy. We conduct experiments to
demonstrate the applications of HoliCity, such as predicting surface
segmentation, normal maps, depth maps, and vanishing points, as well as test
the generalizability of methods trained on HoliCity and other related datasets.
HoliCity is available at https://holicity.io.
Related papers
- Outdoor Scene Extrapolation with Hierarchical Generative Cellular Automata [70.9375320609781]
We aim to generate fine-grained 3D geometry from large-scale sparse LiDAR scans, abundantly captured by autonomous vehicles (AV)
We propose hierarchical Generative Cellular Automata (hGCA), a spatially scalable 3D generative model, which grows geometry with local kernels following, in a coarse-to-fine manner, equipped with a light-weight planner to induce global consistency.
arXiv Detail & Related papers (2024-06-12T14:56:56Z) - CityRefer: Geography-aware 3D Visual Grounding Dataset on City-scale
Point Cloud Data [15.526523262690965]
We introduce the CityRefer dataset for city-level visual grounding.
The dataset consists of 35k natural language descriptions of 3D objects appearing in SensatUrban city scenes and 5k landmarks labels synchronizing with OpenStreetMap.
arXiv Detail & Related papers (2023-10-28T18:05:32Z) - Building3D: An Urban-Scale Dataset and Benchmarks for Learning Roof
Structures from Point Clouds [4.38301148531795]
Existing datasets for 3D modeling mainly focus on common objects such as furniture or cars.
We present a urban-scale dataset consisting of more than 160 thousands buildings along with corresponding point clouds, mesh and wire-frame models, covering 16 cities in Estonia about 998 Km2.
Experimental results indicate that Building3D has challenges of high intra-class variance, data imbalance and large-scale noises.
arXiv Detail & Related papers (2023-07-21T21:38:57Z) - Visual Localization using Imperfect 3D Models from the Internet [54.731309449883284]
This paper studies how imperfections in 3D models affect localization accuracy.
We show that 3D models from the Internet show promise as an easy-to-obtain scene representation.
arXiv Detail & Related papers (2023-04-12T16:15:05Z) - sat2pc: Estimating Point Cloud of Building Roofs from 2D Satellite
Images [1.8884278918443564]
We propose sat2pc, a deep learning architecture that predicts the point of a building roof from a single 2D satellite image.
Our results show that sat2pc was able to outperform existing baselines by at least 18.6%.
arXiv Detail & Related papers (2022-05-25T03:24:40Z) - ImpliCity: City Modeling from Satellite Images with Deep Implicit
Occupancy Fields [20.00737387884824]
ImpliCity is a neural representation of the 3D scene as an implicit, continuous occupancy field, driven by learned embeddings of the point cloud and a stereo pair of ortho-photos.
With image resolution 0.5$,$m, ImpliCity reaches a median height error of $approx,$0.7$,$m and outperforms competing methods.
arXiv Detail & Related papers (2022-01-24T21:40:16Z) - SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point
Clouds [52.624157840253204]
We introduce SensatUrban, an urban-scale UAV photogrammetry point cloud dataset consisting of nearly three billion points collected from three UK cities, covering 7.6 km2.
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
arXiv Detail & Related papers (2022-01-12T14:48:11Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Towards General Purpose Geometry-Preserving Single-View Depth Estimation [1.9573380763700712]
Single-view depth estimation (SVDE) plays a crucial role in scene understanding for AR applications, 3D modeling, and robotics.
Recent works have shown that a successful solution strongly relies on the diversity and volume of training data.
Our work shows that a model trained on this data along with conventional datasets can gain accuracy while predicting correct scene geometry.
arXiv Detail & Related papers (2020-09-25T20:06:13Z) - Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset,
Benchmarks and Challenges [52.624157840253204]
We present an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points.
Our dataset consists of large areas from three UK cities, covering about 7.6 km2 of the city landscape.
We evaluate the performance of state-of-the-art algorithms on our dataset and provide a comprehensive analysis of the results.
arXiv Detail & Related papers (2020-09-07T14:47:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.