TMBuD: A dataset for urban scene building detection
- URL: http://arxiv.org/abs/2110.14590v1
- Date: Wed, 27 Oct 2021 17:08:11 GMT
- Title: TMBuD: A dataset for urban scene building detection
- Authors: Orhei Ciprian, Vert Silviu, Mocofan Muguras, Vasiu Radu
- Abstract summary: This paper introduces a dataset solution, the TMBuD, that is better fitted for image processing on human made structures for urban scene scenarios.
The proposed dataset will allow proper evaluation of salient edges and semantic segmentation of images focusing on the street view perspective of buildings.
The dataset features 160 images of buildings from Timisoara, Romania, with a resolution of 768 x 1024 pixels each.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Building recognition and 3D reconstruction of human made structures in urban
scenarios has become an interesting and actual topic in the image processing
domain. For this research topic the Computer Vision and Augmented Reality areas
intersect for creating a better understanding of the urban scenario for various
topics. In this paper we aim to introduce a dataset solution, the TMBuD, that
is better fitted for image processing on human made structures for urban scene
scenarios. The proposed dataset will allow proper evaluation of salient edges
and semantic segmentation of images focusing on the street view perspective of
buildings. The images that form our dataset offer various street view
perspectives of buildings from urban scenarios, which allows for evaluating
complex algorithms. The dataset features 160 images of buildings from
Timisoara, Romania, with a resolution of 768 x 1024 pixels each.
Related papers
- MaRINeR: Enhancing Novel Views by Matching Rendered Images with Nearby References [49.71130133080821]
MaRINeR is a refinement method that leverages information of a nearby mapping image to improve the rendering of a target viewpoint.
We show improved renderings in quantitative metrics and qualitative examples from both explicit and implicit scene representations.
arXiv Detail & Related papers (2024-07-18T17:50:03Z) - Urban Scene Diffusion through Semantic Occupancy Map [49.20779809250597]
UrbanDiffusion is a 3D diffusion model conditioned on a Bird's-Eye View (BEV) map.
Our model learns the data distribution of scene-level structures within a latent space.
After training on real-world driving datasets, our model can generate a wide range of diverse urban scenes.
arXiv Detail & Related papers (2024-03-18T11:54:35Z) - Sat2Scene: 3D Urban Scene Generation from Satellite Images with Diffusion [77.34078223594686]
We propose a novel architecture for direct 3D scene generation by introducing diffusion models into 3D sparse representations and combining them with neural rendering techniques.
Specifically, our approach generates texture colors at the point level for a given geometry using a 3D diffusion model first, which is then transformed into a scene representation in a feed-forward manner.
Experiments in two city-scale datasets show that our model demonstrates proficiency in generating photo-realistic street-view image sequences and cross-view urban scenes from satellite imagery.
arXiv Detail & Related papers (2024-01-19T16:15:37Z) - Building3D: An Urban-Scale Dataset and Benchmarks for Learning Roof
Structures from Point Clouds [4.38301148531795]
Existing datasets for 3D modeling mainly focus on common objects such as furniture or cars.
We present a urban-scale dataset consisting of more than 160 thousands buildings along with corresponding point clouds, mesh and wire-frame models, covering 16 cities in Estonia about 998 Km2.
Experimental results indicate that Building3D has challenges of high intra-class variance, data imbalance and large-scale noises.
arXiv Detail & Related papers (2023-07-21T21:38:57Z) - UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building
Instance Segmentation [50.52615875873055]
UrbanBIS comprises six real urban scenes, with 2.5 billion points, covering a vast area of 10.78 square kilometers.
UrbanBIS provides semantic-level annotations on a rich set of urban objects, including buildings, vehicles, vegetation, roads, and bridges.
UrbanBIS is the first 3D dataset that introduces fine-grained building sub-categories.
arXiv Detail & Related papers (2023-05-04T08:01:38Z) - Street-View Image Generation from a Bird's-Eye View Layout [95.36869800896335]
Bird's-Eye View (BEV) Perception has received increasing attention in recent years.
Data-driven simulation for autonomous driving has been a focal point of recent research.
We propose BEVGen, a conditional generative model that synthesizes realistic and spatially consistent surrounding images.
arXiv Detail & Related papers (2023-01-11T18:39:34Z) - OmniCity: Omnipotent City Understanding with Multi-level and Multi-view
Images [72.4144257192959]
The paper presents OmniCity, a new dataset for omnipotent city understanding from multi-level and multi-view images.
The dataset contains over 100K pixel-wise annotated images that are well-aligned and collected from 25K geo-locations in New York City.
With the new OmniCity dataset, we provide benchmarks for a variety of tasks including building footprint extraction, height estimation, and building plane/instance/fine-grained segmentation.
arXiv Detail & Related papers (2022-08-01T15:19:25Z) - Re-designing cities with conditional adversarial networks [0.0]
This paper introduces a conditional generative adversarial network to redesign a street-level image of urban scenes.
We introduce a new dataset that comprises aligned street-level images of before and after urban interventions from real-life scenarios.
The trained model shows strong performance in re-modelling cities, outperforming existing methods that apply image-to-image translation in other domains.
arXiv Detail & Related papers (2021-04-08T19:03:34Z) - Holistic Multi-View Building Analysis in the Wild with Projection
Pooling [18.93067906200084]
We address six different classification tasks related to fine-grained building attributes.
Tackling such a remote building analysis problem became possible only recently due to growing large-scale datasets of urban scenes.
We introduce a new benchmarking dataset, consisting of 49426 images (top-view and street-view) of 9674 buildings.
arXiv Detail & Related papers (2020-08-23T13:49:22Z) - Urban2Vec: Incorporating Street View Imagery and POIs for Multi-Modal
Urban Neighborhood Embedding [8.396746290518102]
Urban2Vec is an unsupervised multi-modal framework which incorporates both street view imagery and point-of-interest data.
We show that Urban2Vec can achieve performances better than baseline models and comparable to fully-supervised methods in downstream prediction tasks.
arXiv Detail & Related papers (2020-01-29T21:30:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.