Building Age Estimation: A New Multi-Modal Benchmark Dataset and Community Challenge
- URL: http://arxiv.org/abs/2502.13818v1
- Date: Wed, 19 Feb 2025 15:31:13 GMT
- Title: Building Age Estimation: A New Multi-Modal Benchmark Dataset and Community Challenge
- Authors: Nikolaos Dionelis, Nicolas Longépé, Alessandra Feliciotti, Mattia Marconcini, Devis Peressutti, Nika Oman Kadunc, JaeWan Park, Hagai Raja Sinulingga, Steve Andreas Immanuel, Ba Tran, Caroline Arnold,
- Abstract summary: Estimating the construction year of buildings is of great importance for sustainability.
By using Artificial Intelligence (AI) and recently proposed Transformer models, we are able to estimate the construction epoch of buildings.
- Score: 32.69530674031928
- License:
- Abstract: Estimating the construction year of buildings is of great importance for sustainability. Sustainable buildings minimize energy consumption and are a key part of responsible and sustainable urban planning and development to effectively combat climate change. By using Artificial Intelligence (AI) and recently proposed Transformer models, we are able to estimate the construction epoch of buildings from a multi-modal dataset. In this paper, we introduce a new benchmark multi-modal dataset, i.e. the Map your City Dataset (MyCD), containing top-view Very High Resolution (VHR) images, Earth Observation (EO) multi-spectral data from the Copernicus Sentinel-2 satellite constellation, and street-view images in many different cities in Europe, co-localized with respect to the building under study and labelled with the construction epoch. We assess EO generalization performance on new/ previously unseen cities that have been held-out from training and appear only during inference. In this work, we present the community-based data challenge we organized based on MyCD. The ESA AI4EO Challenge MapYourCity was opened in 2024 for 4 months. Here, we present the Top-4 performing models, and the main evaluation results. During inference, the performance of the models using both all three input modalities and only the two top-view modalities, i.e. without the street-view images, is examined. The evaluation results show that the models are effective and can achieve good performance on this difficult real-world task of estimating the age of buildings, even on previously unseen cities, as well as even using only the two top-view modalities (i.e. VHR and Sentinel-2) during inference.
Related papers
- VecCity: A Taxonomy-guided Library for Map Entity Representation Learning [48.73446321300362]
Map entity representation learning (MapRL) generates versatile and reusable data representations.
We propose a novel taxonomy for MapRL that organizes models based on functional module-such as encoders, pre-training tasks, and downstream tasks.
We present a taxonomy-driven library, VecCity, which offers easy-to-use interfaces for encoding, pre-training, fine-tuning, and evaluation.
arXiv Detail & Related papers (2024-10-31T07:03:46Z) - LiveXiv -- A Multi-Modal Live Benchmark Based on Arxiv Papers Content [62.816876067499415]
We propose LiveXiv: a scalable evolving live benchmark based on scientific ArXiv papers.
LiveXiv accesses domain-specific manuscripts at any given timestamp and proposes to automatically generate visual question-answer pairs.
We benchmark multiple open and proprietary Large Multi-modal Models (LMMs) on the first version of our benchmark, showing its challenging nature and exposing the models true abilities.
arXiv Detail & Related papers (2024-10-14T17:51:23Z) - Identifying every building's function in large-scale urban areas with multi-modality remote-sensing data [5.18540804614798]
This study proposes a semi-supervised framework to identify every building's function in large-scale urban areas.
optical images, building height, and nighttime-light data are collected to describe the morphological attributes of buildings.
Results are evaluated by 20,000 validation points and statistical survey reports from the government.
arXiv Detail & Related papers (2024-05-08T15:32:20Z) - Cross-City Matters: A Multimodal Remote Sensing Benchmark Dataset for
Cross-City Semantic Segmentation using High-Resolution Domain Adaptation
Networks [82.82866901799565]
We build a new set of multimodal remote sensing benchmark datasets (including hyperspectral, multispectral, SAR) for the study purpose of the cross-city semantic segmentation task.
Beyond the single city, we propose a high-resolution domain adaptation network, HighDAN, to promote the AI model's generalization ability from the multi-city environments.
HighDAN is capable of retaining the spatially topological structure of the studied urban scene well in a parallel high-to-low resolution fusion fashion.
arXiv Detail & Related papers (2023-09-26T23:55:39Z) - Unified Data Management and Comprehensive Performance Evaluation for
Urban Spatial-Temporal Prediction [Experiment, Analysis & Benchmark] [78.05103666987655]
This work addresses challenges in accessing and utilizing diverse urban spatial-temporal datasets.
We introduceatomic files, a unified storage format designed for urban spatial-temporal big data, and validate its effectiveness on 40 diverse datasets.
We conduct extensive experiments using diverse models and datasets, establishing a performance leaderboard and identifying promising research directions.
arXiv Detail & Related papers (2023-08-24T16:20:00Z) - Building3D: An Urban-Scale Dataset and Benchmarks for Learning Roof
Structures from Point Clouds [4.38301148531795]
Existing datasets for 3D modeling mainly focus on common objects such as furniture or cars.
We present a urban-scale dataset consisting of more than 160 thousands buildings along with corresponding point clouds, mesh and wire-frame models, covering 16 cities in Estonia about 998 Km2.
Experimental results indicate that Building3D has challenges of high intra-class variance, data imbalance and large-scale noises.
arXiv Detail & Related papers (2023-07-21T21:38:57Z) - 4Seasons: Benchmarking Visual SLAM and Long-Term Localization for
Autonomous Driving in Challenging Conditions [54.59279160621111]
We present a novel visual SLAM and long-term localization benchmark for autonomous driving in challenging conditions based on the large-scale 4Seasons dataset.
The proposed benchmark provides drastic appearance variations caused by seasonal changes and diverse weather and illumination conditions.
We introduce a new unified benchmark for jointly evaluating visual odometry, global place recognition, and map-based visual localization performance.
arXiv Detail & Related papers (2022-12-31T13:52:36Z) - Times Series Forecasting for Urban Building Energy Consumption Based on
Graph Convolutional Network [20.358180125750046]
Building industry accounts for more than 40% of energy consumption in the United States.
UBEM is the foundation to support the design of energy-efficient communities.
Data-driven models integrated engineering or physical knowledge can significantly improve the urban building energy simulation.
arXiv Detail & Related papers (2021-05-27T19:02:04Z) - The SpaceNet Multi-Temporal Urban Development Challenge [5.191792224645409]
Building footprints provide a useful proxy for a great many humanitarian applications.
In this paper we discuss efforts to develop techniques for precise building footprint localization, tracking, and change detection.
The competition centered around a brand new open source dataset of Planet Labs satellite imagery mosaics at 4m resolution.
Winning participants demonstrated impressive performance with the newly developed SpaceNet Change and Object Tracking (SCOT) metric.
arXiv Detail & Related papers (2021-02-23T22:01:22Z) - Building Footprint Generation by IntegratingConvolution Neural Network
with Feature PairwiseConditional Random Field (FPCRF) [21.698236040666675]
Building footprint maps are vital to many remote sensing applications, such as 3D building modeling, urban planning, and disaster management.
In this work, an end-to-end building footprint generation approach that integrates convolution neural network (CNN) and graph model is proposed.
arXiv Detail & Related papers (2020-02-11T18:51:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.