Spot: A Natural Language Interface for Geospatial Searches in OSM
- URL: http://arxiv.org/abs/2311.08093v1
- Date: Tue, 14 Nov 2023 11:35:09 GMT
- Title: Spot: A Natural Language Interface for Geospatial Searches in OSM
- Authors: Lynn Khellaf, Ipek Baris Schlicht, Julia Bayer, Ruben Bouwmeester,
Tilman Mira{\ss} and Tilman Wagner
- Abstract summary: Spot is a user-friendly natural language interface for querying OpenStreetMap data.
It extracts relevant information from user-input sentences and displays candidate locations matching the descriptions on a map.
All code and generated data is available as an open-source repository.
- Score: 0.9320657506524149
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Investigative journalists and fact-checkers have found OpenStreetMap (OSM) to
be an invaluable resource for their work due to its extensive coverage and
intricate details of various locations, which play a crucial role in
investigating news scenes. Despite its value, OSM's complexity presents
considerable accessibility and usability challenges, especially for those
without a technical background. To address this, we introduce 'Spot', a
user-friendly natural language interface for querying OSM data. Spot utilizes a
semantic mapping from natural language to OSM tags, leveraging artificially
generated sentence queries and a T5 transformer. This approach enables Spot to
extract relevant information from user-input sentences and display candidate
locations matching the descriptions on a map. To foster collaboration and
future advancement, all code and generated data is available as an open-source
repository.
Related papers
- SPOT: Bridging Natural Language and Geospatial Search for Investigative Journalists [0.8796261172196742]
SPOT, an open source natural language interface, makes OSM's rich, tag-based geographic data more accessible through intuitive scene descriptions.<n>It addresses real-world challenges such as hallucinations in model output, inconsistencies in OSM tagging, and the noisy nature of user input.<n>To our knowledge, SPOT is the first system to achieve reliable natural language access to OSM data at this level of accuracy.
arXiv Detail & Related papers (2025-06-16T07:55:44Z) - Grounding-MD: Grounded Video-language Pre-training for Open-World Moment Detection [67.70328796057466]
Grounding-MD is an innovative, grounded video-language pre-training framework tailored for open-world moment detection.
Our framework incorporates an arbitrary number of open-ended natural language queries through a structured prompt mechanism.
Grounding-MD demonstrates exceptional semantic representation learning capabilities, effectively handling diverse and complex query conditions.
arXiv Detail & Related papers (2025-04-20T09:54:25Z) - FindAnything: Open-Vocabulary and Object-Centric Mapping for Robot Exploration in Any Environment [16.987872206495897]
FindAnything is an open-world mapping framework that incorporates vision-language information into dense volumetric submaps.
Our system is the first of its kind to be deployed on resource-constrained devices, such as MAVs.
arXiv Detail & Related papers (2025-04-11T15:12:05Z) - OmniGeo: Towards a Multimodal Large Language Models for Geospatial Artificial Intelligence [51.0456395687016]
multimodal large language models (LLMs) have opened new frontiers in artificial intelligence.
We propose a MLLM (OmniGeo) tailored to geospatial applications.
By combining the strengths of natural language understanding and spatial reasoning, our model enhances the ability of instruction following and the accuracy of GeoAI systems.
arXiv Detail & Related papers (2025-03-20T16:45:48Z) - Where am I? Cross-View Geo-localization with Natural Language Descriptions [19.041103952139178]
Cross-view geo-localization identifies the locations of street-view images by matching them with geo-tagged satellite images or OSM.
We introduce a novel task for cross-view geo-localization with natural language descriptions, which aims to retrieve corresponding satellite images or OSM database based on scene text descriptions.
arXiv Detail & Related papers (2024-12-22T13:13:10Z) - Where on Earth Do Users Say They Are?: Geo-Entity Linking for Noisy Multilingual User Input [2.516307239032451]
We present a method which represents real-world locations as averaged embeddings from labeled user-input location names.
We show that our approach improves geo-entity linking on a global and multilingual social media dataset.
arXiv Detail & Related papers (2024-04-29T15:18:33Z) - VLLMs Provide Better Context for Emotion Understanding Through Common Sense Reasoning [66.23296689828152]
We leverage the capabilities of Vision-and-Large-Language Models to enhance in-context emotion classification.
In the first stage, we propose prompting VLLMs to generate descriptions in natural language of the subject's apparent emotion.
In the second stage, the descriptions are used as contextual information and, along with the image input, are used to train a transformer-based architecture.
arXiv Detail & Related papers (2024-04-10T15:09:15Z) - Natural Language Interfaces for Tabular Data Querying and Visualization: A Survey [30.836162812277085]
The rise of large language models (LLMs) has further advanced this field, opening new avenues for natural language processing techniques.
We introduce the fundamental concepts and techniques underlying these interfaces with a particular emphasis on semantic parsing.
This includes a deep dive into the influence of LLMs, highlighting their strengths, limitations, and potential for future improvements.
arXiv Detail & Related papers (2023-10-27T05:01:20Z) - GeoLLM: Extracting Geospatial Knowledge from Large Language Models [49.20315582673223]
We present GeoLLM, a novel method that can effectively extract geospatial knowledge from large language models.
We demonstrate the utility of our approach across multiple tasks of central interest to the international community, including the measurement of population density and economic livelihoods.
Our experiments reveal that LLMs are remarkably sample-efficient, rich in geospatial information, and robust across the globe.
arXiv Detail & Related papers (2023-10-10T00:03:23Z) - Chatmap : Large Language Model Interaction with Cartographic Data [0.0]
OpenStreetMap (OSM) is the most ambitious open-source global initiative offering detailed urban and rural geographic data.
In this study, we demonstrate the proof of concept and details of the process of fine-tuning a relatively small scale (1B parameters) Large Language Models (LLMs) with a relatively small artificial dataset curated by a more capable teacher model.
The study aims to provide an initial guideline for such generative artificial intelligence (AI) adaptations and demonstrate early signs of useful emerging abilities in this context.
arXiv Detail & Related papers (2023-09-28T15:32:36Z) - MGeo: Multi-Modal Geographic Pre-Training Method [49.78466122982627]
We propose a novel query-POI matching method Multi-modal Geographic language model (MGeo)
MGeo represents GC as a new modality and is able to fully extract multi-modal correlations for accurate query-POI matching.
Our proposed multi-modal pre-training method can significantly improve the query-POI matching capability of generic PTMs.
arXiv Detail & Related papers (2023-01-11T03:05:12Z) - Explaining Patterns in Data with Language Models via Interpretable
Autoprompting [143.4162028260874]
We introduce interpretable autoprompting (iPrompt), an algorithm that generates a natural-language string explaining the data.
iPrompt can yield meaningful insights by accurately finding groundtruth dataset descriptions.
Experiments with an fMRI dataset show the potential for iPrompt to aid in scientific discovery.
arXiv Detail & Related papers (2022-10-04T18:32:14Z) - GeoVectors: A Linked Open Corpus of OpenStreetMap Embeddings on World
Scale [1.933681537640272]
This paper presents Geos - a unique, comprehensive world-scale linked corpus of OSM entity embeddings.
The Geos corpus captures semantic and geographic dimensions of OSM entities and makes them accessible to machine learning algorithms.
We provide a SPARQL endpoint - a semantic interface that offers direct access to semantic and latent representations of geographic entities in OSM.
arXiv Detail & Related papers (2021-08-30T10:00:34Z) - Towards Natural Language Question Answering over Earth Observation
Linked Data using Attention-based Neural Machine Translation [0.0]
This paper seeks to study and analyze the use of RNN-based neural machine translation with attention for transforming natural language questions into GeoSPARQL queries.
A dataset consisting of mappings from natural language questions to GeoSPARQL queries over the Corine Land Cover(CLC) Linked Data has been created to train and validate the deep neural network.
arXiv Detail & Related papers (2021-01-23T06:12:20Z) - OpenStreetMap: Challenges and Opportunities in Machine Learning and
Remote Sensing [66.23463054467653]
We present a review of recent methods based on machine learning to improve and use OpenStreetMap data.
We believe that OSM can change the way we interpret remote sensing data and that the synergy with machine learning can scale participatory map making.
arXiv Detail & Related papers (2020-07-13T09:58:14Z) - Language Guided Networks for Cross-modal Moment Retrieval [66.49445903955777]
Cross-modal moment retrieval aims to localize a temporal segment from an untrimmed video described by a natural language query.
Existing methods independently extract the features of videos and sentences.
We present Language Guided Networks (LGN), a new framework that leverages the sentence embedding to guide the whole process of moment retrieval.
arXiv Detail & Related papers (2020-06-18T12:08:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.