A Copy Mechanism for Handling Knowledge Base Elements in SPARQL Neural
Machine Translation
- URL: http://arxiv.org/abs/2211.10271v1
- Date: Fri, 18 Nov 2022 14:56:35 GMT
- Title: A Copy Mechanism for Handling Knowledge Base Elements in SPARQL Neural
Machine Translation
- Authors: Rose Hirigoyen, Amal Zouaq and Samuel Reyd
- Abstract summary: We propose to integrate a copy mechanism for neural SPARQL query generation as a way to tackle this issue.
We illustrate our proposal by adding a copy layer and a dynamic knowledge base vocabulary to two Seq2Seq architectures (CNNs and Transformers)
This layer makes the models copy KB elements directly from the questions, instead of generating them.
We evaluate our approach on state-of-the-art datasets, including datasets referencing unknown KB elements and measure the accuracy of the copy-augmented architectures.
- Score: 2.9134135167113433
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural Machine Translation (NMT) models from English to SPARQL are a
promising development for SPARQL query generation. However, current
architectures are unable to integrate the knowledge base (KB) schema and handle
questions on knowledge resources, classes, and properties unseen during
training, rendering them unusable outside the scope of topics covered in the
training set. Inspired by the performance gains in natural language processing
tasks, we propose to integrate a copy mechanism for neural SPARQL query
generation as a way to tackle this issue. We illustrate our proposal by adding
a copy layer and a dynamic knowledge base vocabulary to two Seq2Seq
architectures (CNNs and Transformers). This layer makes the models copy KB
elements directly from the questions, instead of generating them. We evaluate
our approach on state-of-the-art datasets, including datasets referencing
unknown KB elements and measure the accuracy of the copy-augmented
architectures. Our results show a considerable increase in performance on all
datasets compared to non-copy architectures.
Related papers
- Two are better than one: Context window extension with multi-grained self-injection [111.1376461868317]
SharedLLM is a novel approach grounded in the design philosophy of multi-grained context compression and query-aware information retrieval.
We introduce a specialized tree-style data structure to efficiently encode, store and retrieve multi-grained contextual information for text chunks.
arXiv Detail & Related papers (2024-10-25T06:08:59Z) - Less is More: Making Smaller Language Models Competent Subgraph Retrievers for Multi-hop KGQA [51.3033125256716]
We model the subgraph retrieval task as a conditional generation task handled by small language models.
Our base generative subgraph retrieval model, consisting of only 220M parameters, competitive retrieval performance compared to state-of-the-art models.
Our largest 3B model, when plugged with an LLM reader, sets new SOTA end-to-end performance on both the WebQSP and CWQ benchmarks.
arXiv Detail & Related papers (2024-10-08T15:22:36Z) - A Comprehensive Evaluation of Neural SPARQL Query Generation from
Natural Language Questions [2.5782420501870296]
In recent years, the field of neural machine translation (NMT) for SPARQL query generation has witnessed significant growth.
This paper presents various experiments that replicate and expand upon recent NMT-based SPARQL generation studies.
arXiv Detail & Related papers (2023-04-16T13:12:26Z) - Knowledge Graph Question Answering via SPARQL Silhouette Generation [18.391235417154498]
Knowledge Graph Question Answering (KGQA) has become a prominent area in natural language processing.
We propose a modular two-stage neural architecture to solve the KGQA task.
We show that our method can achieve reasonable performance improving the state-of-art by a margin of 3.72% F1 for the LC-QuAD-1 dataset.
arXiv Detail & Related papers (2021-09-06T14:55:37Z) - SPBERT: Pre-training BERT on SPARQL Queries for End-to-end Question
Answering over Knowledge Graphs [1.1775939485654976]
SPBERT is a Transformer-based language model pre-trained on massive SPARQL query logs.
We investigate how SPBERT and encoder-decoder architecture can be adapted for Knowledge-based QA corpora.
arXiv Detail & Related papers (2021-06-18T08:39:26Z) - Exploring Sequence-to-Sequence Models for SPARQL Pattern Composition [0.5639451539396457]
A booming amount of information is continuously added to the Internet as structured and unstructured data, feeding knowledge bases such as DBpedia and Wikidata.
The aim of Question Answering systems is to allow lay users to access such data using natural language without needing to write formal queries.
We show that sequence-to-sequence models are a viable and promising option to transform long utterances into complex SPARQL queries.
arXiv Detail & Related papers (2020-10-21T11:12:01Z) - Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks [133.93803565077337]
retrieval-augmented generation models combine pre-trained parametric and non-parametric memory for language generation.
We show that RAG models generate more specific, diverse and factual language than a state-of-the-art parametric-only seq2seq baseline.
arXiv Detail & Related papers (2020-05-22T21:34:34Z) - Probing Linguistic Features of Sentence-Level Representations in Neural
Relation Extraction [80.38130122127882]
We introduce 14 probing tasks targeting linguistic properties relevant to neural relation extraction (RE)
We use them to study representations learned by more than 40 different encoder architecture and linguistic feature combinations trained on two datasets.
We find that the bias induced by the architecture and the inclusion of linguistic features are clearly expressed in the probing task performance.
arXiv Detail & Related papers (2020-04-17T09:17:40Z) - Conversational Question Reformulation via Sequence-to-Sequence
Architectures and Pretrained Language Models [56.268862325167575]
This paper presents an empirical study of conversational question reformulation (CQR) with sequence-to-sequence architectures and pretrained language models (PLMs)
We leverage PLMs to address the strong token-to-token independence assumption made in the common objective, maximum likelihood estimation, for the CQR task.
We evaluate fine-tuned PLMs on the recently-introduced CANARD dataset as an in-domain task and validate the models using data from the TREC 2019 CAsT Track as an out-domain task.
arXiv Detail & Related papers (2020-04-04T11:07:54Z) - Hierarchical Conditional Relation Networks for Video Question Answering [62.1146543269993]
We introduce a general-purpose reusable neural unit called Conditional Relation Network (CRN)
CRN serves as a building block to construct more sophisticated structures for representation and reasoning over video.
Our evaluations on well-known datasets achieved new SoTA results, demonstrating the impact of building a general-purpose reasoning unit on complex domains such as VideoQA.
arXiv Detail & Related papers (2020-02-25T07:00:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.