A citizen science toolkit to collect human perceptions of urban environments using open street view images
- URL: http://arxiv.org/abs/2403.00174v4
- Date: Mon, 04 Nov 2024 11:38:49 GMT
- Title: A citizen science toolkit to collect human perceptions of urban environments using open street view images
- Authors: Matthew Danish, SM Labib, Britta Ricker, Marco Helbich,
- Abstract summary: Street View Imagery (SVI) is a valuable data source for studies (e.g., environmental assessments, green space identification or land cover classification)
Open SVI datasets are readily available from less restrictive sources, such as Mapillary.
We present an efficient method for automated downloading, processing, cropping, and filtering open SVI.
- Score: 0.20999222360659603
- License:
- Abstract: Street View Imagery (SVI) is a valuable data source for studies (e.g., environmental assessments, green space identification or land cover classification). While commercial SVI is available, such providers commonly restrict copying or reuse in ways necessary for research. Open SVI datasets are readily available from less restrictive sources, such as Mapillary, but due to the heterogeneity of the images, these require substantial preprocessing, filtering, and careful quality checks. We present an efficient method for automated downloading, processing, cropping, and filtering open SVI, to be used in a survey of human perceptions of the streets portrayed in these images. We demonstrate our open-source reusable SVI preparation and smartphone-friendly perception-survey software with Amsterdam (Netherlands) as the case study. Using a citizen science approach, we collected from 331 people 22,637 ratings about their perceptions for various criteria. We have published our software in a public repository for future re-use and reproducibility.
Related papers
- Bird's-Eye View to Street-View: A Survey [16.90516098120805]
We screened 20 recent research papers to review the state-of-the-art of how street-view images are synthesized from their corresponding satellite counterparts.
Main findings are: (i) novel deep learning techniques are required for synthesizing more realistic and accurate street-view images; (ii) more datasets need to be collected for public usage; and (iii) more specific evaluation metrics need to be investigated for evaluating the generated images appropriately.
arXiv Detail & Related papers (2024-05-14T21:01:12Z) - The Adversarial AI-Art: Understanding, Generation, Detection, and Benchmarking [47.08666835021915]
We present a systematic attempt at understanding and detecting AI-generated images (AI-art) in adversarial scenarios.
The dataset, named ARIA, contains over 140K images in five categories: artworks (painting), social media images, news photos, disaster scenes, and anime pictures.
arXiv Detail & Related papers (2024-04-22T21:00:13Z) - Cross-Modal Learning of Housing Quality in Amsterdam [7.316396716394438]
We test data and models for the recognition of housing quality in Amsterdam from ground-level and aerial imagery.
For ground-level images we compare Google StreetView (GSV) to Flickr images.
Our results show that GSV predicts the most accurate building quality scores, approximately 30% better than using only aerial images.
arXiv Detail & Related papers (2024-03-13T19:11:58Z) - To use or not to use proprietary street view images in (health and place) research? That is the question [0.20999222360659603]
This article questions the current practices in using Google Street View images from a European viewpoint.
Our concern lies with Google's terms of service, which restrict bulk image downloads and the generation of street view image-based indices.
arXiv Detail & Related papers (2024-02-18T08:26:22Z) - Open-source data pipeline for street-view images: a case study on
community mobility during COVID-19 pandemic [0.9423257767158634]
Street View Images (SVI) are a common source of valuable data for researchers.
Google Street View images are collected infrequently, making temporal analysis challenging.
This study demonstrates the feasibility and value of collecting and using SVI for research purposes beyond what is possible with currently available SVI data.
arXiv Detail & Related papers (2024-01-23T20:56:16Z) - Raising the Bar of AI-generated Image Detection with CLIP [50.345365081177555]
The aim of this work is to explore the potential of pre-trained vision-language models (VLMs) for universal detection of AI-generated images.
We develop a lightweight detection strategy based on CLIP features and study its performance in a wide variety of challenging scenarios.
arXiv Detail & Related papers (2023-11-30T21:11:20Z) - NPF-200: A Multi-Modal Eye Fixation Dataset and Method for
Non-Photorealistic Videos [51.409547544747284]
NPF-200 is the first large-scale multi-modal dataset of purely non-photorealistic videos with eye fixations.
We conduct a series of analyses to gain deeper insights into this task.
We propose a universal frequency-aware multi-modal non-photorealistic saliency detection model called NPSNet.
arXiv Detail & Related papers (2023-08-23T14:25:22Z) - A Comprehensive Survey of Transformers for Computer Vision [3.1761172592339375]
Vision Transformers (ViTs) are used to various computer vision applications (CV)
This survey is the first of its kind on ViTs for CVs to the best of our knowledge.
CV applications include image classification, object detection, image segmentation, image compression, image super-resolution, image denoising, and anomaly detection.
arXiv Detail & Related papers (2022-11-11T05:11:03Z) - Perceptual Quality Assessment of Omnidirectional Images [81.76416696753947]
We first establish an omnidirectional IQA (OIQA) database, which includes 16 source images and 320 distorted images degraded by 4 commonly encountered distortion types.
Then a subjective quality evaluation study is conducted on the OIQA database in the VR environment.
The original and distorted omnidirectional images, subjective quality ratings, and the head and eye movement data together constitute the OIQA database.
arXiv Detail & Related papers (2022-07-06T13:40:38Z) - APES: Audiovisual Person Search in Untrimmed Video [87.4124877066541]
We present the Audiovisual Person Search dataset (APES)
APES contains over 1.9K identities labeled along 36 hours of video.
A key property of APES is that it includes dense temporal annotations that link faces to speech segments of the same identity.
arXiv Detail & Related papers (2021-06-03T08:16:42Z) - Lighting the Darkness in the Deep Learning Era [118.35081853500411]
Low-light image enhancement (LLIE) aims at improving the perception or interpretability of an image captured in an environment with poor illumination.
Recent advances in this area are dominated by deep learning-based solutions.
We provide a comprehensive survey to cover various aspects ranging from algorithm taxonomy to unsolved open issues.
arXiv Detail & Related papers (2021-04-21T19:12:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.