Re-designing cities with conditional adversarial networks
- URL: http://arxiv.org/abs/2104.04013v1
- Date: Thu, 8 Apr 2021 19:03:34 GMT
- Title: Re-designing cities with conditional adversarial networks
- Authors: Mohamed R. Ibrahim, James Haworth, Nicola Christie
- Abstract summary: This paper introduces a conditional generative adversarial network to redesign a street-level image of urban scenes.
We introduce a new dataset that comprises aligned street-level images of before and after urban interventions from real-life scenarios.
The trained model shows strong performance in re-modelling cities, outperforming existing methods that apply image-to-image translation in other domains.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper introduces a conditional generative adversarial network to
redesign a street-level image of urban scenes by generating 1) an urban
intervention policy, 2) an attention map that localises where intervention is
needed, 3) a high-resolution street-level image (1024 X 1024 or 1536 X1536)
after implementing the intervention. We also introduce a new dataset that
comprises aligned street-level images of before and after urban interventions
from real-life scenarios that make this research possible. The introduced
method has been trained on different ranges of urban interventions applied to
realistic images. The trained model shows strong performance in re-modelling
cities, outperforming existing methods that apply image-to-image translation in
other domains that is computed in a single GPU. This research opens the door
for machine intelligence to play a role in re-thinking and re-designing the
different attributes of cities based on adversarial learning, going beyond the
mainstream of facial landmarks manipulation or image synthesis from semantic
segmentation.
Related papers
- MaRINeR: Enhancing Novel Views by Matching Rendered Images with Nearby References [49.71130133080821]
MaRINeR is a refinement method that leverages information of a nearby mapping image to improve the rendering of a target viewpoint.
We show improved renderings in quantitative metrics and qualitative examples from both explicit and implicit scene representations.
arXiv Detail & Related papers (2024-07-18T17:50:03Z) - AddressCLIP: Empowering Vision-Language Models for City-wide Image Address Localization [57.34659640776723]
We propose an end-to-end framework named AddressCLIP to solve the problem with more semantics.
We have built three datasets from Pittsburgh and San Francisco on different scales specifically for the IAL problem.
arXiv Detail & Related papers (2024-07-11T03:18:53Z) - Multi-scale Semantic Prior Features Guided Deep Neural Network for Urban Street-view Image [1.4473649585131072]
This paper presents a novel Deep Neural Network (DNN) for inpainting street-view images.
A semantic prior prompter is introduced to learn rich semantic priors from large pre-trained model.
Experiments on Apolloscapes and Cityscapes datasets demonstrate better performance than state-of-the-art methods.
arXiv Detail & Related papers (2024-05-17T03:02:18Z) - Urban Scene Diffusion through Semantic Occupancy Map [49.20779809250597]
UrbanDiffusion is a 3D diffusion model conditioned on a Bird's-Eye View (BEV) map.
Our model learns the data distribution of scene-level structures within a latent space.
After training on real-world driving datasets, our model can generate a wide range of diverse urban scenes.
arXiv Detail & Related papers (2024-03-18T11:54:35Z) - Sat2Scene: 3D Urban Scene Generation from Satellite Images with Diffusion [77.34078223594686]
We propose a novel architecture for direct 3D scene generation by introducing diffusion models into 3D sparse representations and combining them with neural rendering techniques.
Specifically, our approach generates texture colors at the point level for a given geometry using a 3D diffusion model first, which is then transformed into a scene representation in a feed-forward manner.
Experiments in two city-scale datasets show that our model demonstrates proficiency in generating photo-realistic street-view image sequences and cross-view urban scenes from satellite imagery.
arXiv Detail & Related papers (2024-01-19T16:15:37Z) - A shape-based heuristic for the detection of urban block artifacts in
street networks [0.0]
Street networks are ubiquitous components of cities, guiding their development and enabling movement from place to place.
Their graph representation is often designed primarily for transportation purposes.
This representation is less suitable for other use cases where transportation networks need to be simplified.
arXiv Detail & Related papers (2023-09-01T13:11:35Z) - Mitigating Urban-Rural Disparities in Contrastive Representation Learning with Satellite Imagery [19.93324644519412]
We consider the risk of urban-rural disparities in identification of land-cover features.
We propose fair dense representation with contrastive learning (FairDCL) as a method for de-biasing the multi-level latent space of convolution neural network models.
The obtained image representation mitigates downstream urban-rural prediction disparities and outperforms state-of-the-art baselines on real-world satellite images.
arXiv Detail & Related papers (2022-11-16T04:59:46Z) - TMBuD: A dataset for urban scene building detection [0.0]
This paper introduces a dataset solution, the TMBuD, that is better fitted for image processing on human made structures for urban scene scenarios.
The proposed dataset will allow proper evaluation of salient edges and semantic segmentation of images focusing on the street view perspective of buildings.
The dataset features 160 images of buildings from Timisoara, Romania, with a resolution of 768 x 1024 pixels each.
arXiv Detail & Related papers (2021-10-27T17:08:11Z) - Cross-Descriptor Visual Localization and Mapping [81.16435356103133]
Visual localization and mapping is the key technology underlying the majority of Mixed Reality and robotics systems.
We present three novel scenarios for localization and mapping which require the continuous update of feature representations.
Our data-driven approach is agnostic to the feature descriptor type, has low computational requirements, and scales linearly with the number of description algorithms.
arXiv Detail & Related papers (2020-12-02T18:19:51Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - Urban2Vec: Incorporating Street View Imagery and POIs for Multi-Modal
Urban Neighborhood Embedding [8.396746290518102]
Urban2Vec is an unsupervised multi-modal framework which incorporates both street view imagery and point-of-interest data.
We show that Urban2Vec can achieve performances better than baseline models and comparable to fully-supervised methods in downstream prediction tasks.
arXiv Detail & Related papers (2020-01-29T21:30:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.