Towards Natural Language Question Answering over Earth Observation
Linked Data using Attention-based Neural Machine Translation
- URL: http://arxiv.org/abs/2101.09427v1
- Date: Sat, 23 Jan 2021 06:12:20 GMT
- Title: Towards Natural Language Question Answering over Earth Observation
Linked Data using Attention-based Neural Machine Translation
- Authors: Abhishek V. Potnis, Rajat C. Shinde, Surya S. Durbha
- Abstract summary: This paper seeks to study and analyze the use of RNN-based neural machine translation with attention for transforming natural language questions into GeoSPARQL queries.
A dataset consisting of mappings from natural language questions to GeoSPARQL queries over the Corine Land Cover(CLC) Linked Data has been created to train and validate the deep neural network.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With an increase in Geospatial Linked Open Data being adopted and published
over the web, there is a need to develop intuitive interfaces and systems for
seamless and efficient exploratory analysis of such rich heterogeneous
multi-modal datasets. This work is geared towards improving the exploration
process of Earth Observation (EO) Linked Data by developing a natural language
interface to facilitate querying. Questions asked over Earth Observation Linked
Data have an inherent spatio-temporal dimension and can be represented using
GeoSPARQL. This paper seeks to study and analyze the use of RNN-based neural
machine translation with attention for transforming natural language questions
into GeoSPARQL queries. Specifically, it aims to assess the feasibility of a
neural approach for identifying and mapping spatial predicates in natural
language to GeoSPARQL's topology vocabulary extension including - Egenhofer and
RCC8 relations. The queries can then be executed over a triple store to yield
answers for the natural language questions. A dataset consisting of mappings
from natural language questions to GeoSPARQL queries over the Corine Land
Cover(CLC) Linked Data has been created to train and validate the deep neural
network. From our experiments, it is evident that neural machine translation
with attention is a promising approach for the task of translating spatial
predicates in natural language questions to GeoSPARQL queries.
Related papers
- GeoReasoner: Reasoning On Geospatially Grounded Context For Natural Language Understanding [0.32885740436059047]
GeoReasoner is a language model capable of reasoning on geospatially grounded natural language.
It first leverages Large Language Models to generate a comprehensive location description based on linguistic inferences and distance information.
It also encodes direction and distance information into spatial embedding via treating them as pseudo-sentences.
arXiv Detail & Related papers (2024-08-21T06:35:21Z) - Into the Unknown: Generating Geospatial Descriptions for New Environments [18.736071151303726]
Rendezvous task requires reasoning over allocentric spatial relationships.
Using opensource descriptions paired with coordinates (e.g., Wikipedia) provides training data but suffers from limited spatially-oriented text.
We propose a large-scale augmentation method for generating high-quality synthetic data for new environments.
arXiv Detail & Related papers (2024-06-28T14:56:21Z) - Geode: A Zero-shot Geospatial Question-Answering Agent with Explicit Reasoning and Precise Spatio-Temporal Retrieval [0.0]
We introduce a pioneering system designed to tackle zero-shot geospatial question-answering tasks with high precision.
Our approach represents a significant improvement in addressing the limitations of current large language models.
arXiv Detail & Related papers (2024-06-26T21:59:54Z) - Natural Language Processing for Dialects of a Language: A Survey [56.93337350526933]
State-of-the-art natural language processing (NLP) models are trained on massive training corpora, and report a superlative performance on evaluation datasets.
This survey delves into an important attribute of these datasets: the dialect of a language.
Motivated by the performance degradation of NLP models for dialectic datasets and its implications for the equity of language technologies, we survey past research in NLP for dialects in terms of datasets, and approaches.
arXiv Detail & Related papers (2024-01-11T03:04:38Z) - Towards Natural Language-Guided Drones: GeoText-1652 Benchmark with Spatial Relation Matching [60.645802236700035]
Navigating drones through natural language commands remains challenging due to the dearth of accessible multi-modal datasets.
We introduce GeoText-1652, a new natural language-guided geo-localization benchmark.
This dataset is systematically constructed through an interactive human-computer process.
arXiv Detail & Related papers (2023-11-21T17:52:30Z) - GeoLM: Empowering Language Models for Geospatially Grounded Language
Understanding [45.36562604939258]
This paper introduces GeoLM, a language model that enhances the understanding of geo-entities in natural language.
We demonstrate that GeoLM exhibits promising capabilities in supporting toponym recognition, toponym linking, relation extraction, and geo-entity typing.
arXiv Detail & Related papers (2023-10-23T01:20:01Z) - Core Building Blocks: Next Gen Geo Spatial GPT Application [0.0]
This paper introduces MapGPT, which aims to bridge the gap between natural language understanding and spatial data analysis.
MapGPT enables more accurate and contextually aware responses to location-based queries.
arXiv Detail & Related papers (2023-10-17T06:59:31Z) - MGeo: Multi-Modal Geographic Pre-Training Method [49.78466122982627]
We propose a novel query-POI matching method Multi-modal Geographic language model (MGeo)
MGeo represents GC as a new modality and is able to fully extract multi-modal correlations for accurate query-POI matching.
Our proposed multi-modal pre-training method can significantly improve the query-POI matching capability of generic PTMs.
arXiv Detail & Related papers (2023-01-11T03:05:12Z) - Explaining Patterns in Data with Language Models via Interpretable
Autoprompting [143.4162028260874]
We introduce interpretable autoprompting (iPrompt), an algorithm that generates a natural-language string explaining the data.
iPrompt can yield meaningful insights by accurately finding groundtruth dataset descriptions.
Experiments with an fMRI dataset show the potential for iPrompt to aid in scientific discovery.
arXiv Detail & Related papers (2022-10-04T18:32:14Z) - GATE: Graph Attention Transformer Encoder for Cross-lingual Relation and
Event Extraction [107.8262586956778]
We introduce graph convolutional networks (GCNs) with universal dependency parses to learn language-agnostic sentence representations.
GCNs struggle to model words with long-range dependencies or are not directly connected in the dependency tree.
We propose to utilize the self-attention mechanism to learn the dependencies between words with different syntactic distances.
arXiv Detail & Related papers (2020-10-06T20:30:35Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.