AutoQGS: Auto-Prompt for Low-Resource Knowledge-based Question
Generation from SPARQL
- URL: http://arxiv.org/abs/2208.12461v1
- Date: Fri, 26 Aug 2022 06:53:46 GMT
- Title: AutoQGS: Auto-Prompt for Low-Resource Knowledge-based Question
Generation from SPARQL
- Authors: Guanming Xiong, Junwei Bao, Wen Zhao, Youzheng Wu, Xiaodong He
- Abstract summary: This study investigates the task of knowledge-based question generation (KBQG)
Conventional KBQG works generated questions from fact triples in the knowledge graph, which could not express complex operations like aggregation and comparison in SPARQL.
We propose an auto-prompter trained on large-scale unsupervised data to rephrase SPARQL into NL description.
- Score: 18.019353543946913
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This study investigates the task of knowledge-based question generation
(KBQG). Conventional KBQG works generated questions from fact triples in the
knowledge graph, which could not express complex operations like aggregation
and comparison in SPARQL. Moreover, due to the costly annotation of large-scale
SPARQL-question pairs, KBQG from SPARQL under low-resource scenarios urgently
needs to be explored. Recently, since the generative pre-trained language
models (PLMs) typically trained in natural language (NL)-to-NL paradigm have
been proven effective for low-resource generation, e.g., T5 and BART, how to
effectively utilize them to generate NL-question from non-NL SPARQL is
challenging. To address these challenges, AutoQGS, an auto-prompt approach for
low-resource KBQG from SPARQL, is proposed. Firstly, we put forward to generate
questions directly from SPARQL for the KBQG task to handle complex operations.
Secondly, we propose an auto-prompter trained on large-scale unsupervised data
to rephrase SPARQL into NL description, smoothing the low-resource
transformation from non-NL SPARQL to NL question with PLMs. Experimental
results on the WebQuestionsSP, ComlexWebQuestions 1.1, and PathQuestions show
that our model achieves state-of-the-art performance, especially in
low-resource settings. Furthermore, a corpus of 330k factoid complex
question-SPARQL pairs is generated for further KBQG research.
Related papers
- Effective Instruction Parsing Plugin for Complex Logical Query Answering on Knowledge Graphs [51.33342412699939]
Knowledge Graph Query Embedding (KGQE) aims to embed First-Order Logic (FOL) queries in a low-dimensional KG space for complex reasoning over incomplete KGs.
Recent studies integrate various external information (such as entity types and relation context) to better capture the logical semantics of FOL queries.
We propose an effective Query Instruction Parsing (QIPP) that captures latent query patterns from code-like query instructions.
arXiv Detail & Related papers (2024-10-27T03:18:52Z) - Less is More: Making Smaller Language Models Competent Subgraph Retrievers for Multi-hop KGQA [51.3033125256716]
We model the subgraph retrieval task as a conditional generation task handled by small language models.
Our base generative subgraph retrieval model, consisting of only 220M parameters, competitive retrieval performance compared to state-of-the-art models.
Our largest 3B model, when plugged with an LLM reader, sets new SOTA end-to-end performance on both the WebQSP and CWQ benchmarks.
arXiv Detail & Related papers (2024-10-08T15:22:36Z) - SPARQL Generation: an analysis on fine-tuning OpenLLaMA for Question
Answering over a Life Science Knowledge Graph [0.0]
We evaluate strategies for fine-tuning the OpenLlama LLM for question answering over life science knowledge graphs.
We propose an end-to-end data augmentation approach for extending a set of existing queries over a given knowledge graph.
We also investigate the role of semantic "clues" in the queries, such as meaningful variable names and inline comments.
arXiv Detail & Related papers (2024-02-07T07:24:01Z) - Leveraging LLMs in Scholarly Knowledge Graph Question Answering [7.951847862547378]
KGQA answers natural language questions by leveraging a large language model (LLM)
Our system achieves an F1 score of 99.0% on SciQA - one of the Scholarly Knowledge Graph Question Answering challenge benchmarks.
arXiv Detail & Related papers (2023-11-16T12:13:49Z) - In-Context Learning for Knowledge Base Question Answering for Unmanned
Systems based on Large Language Models [43.642717344626355]
We focus on the CCKS2023 Competition of Question Answering with Knowledge Graph Inference for Unmanned Systems.
Inspired by the recent success of large language models (LLMs) like ChatGPT and GPT-3 in many QA tasks, we propose a ChatGPT-based Cypher Query Language (CQL) generation framework.
With our ChatGPT-based CQL generation framework, we achieved the second place in the CCKS 2023 Question Answering with Knowledge Graph Inference for Unmanned Systems competition.
arXiv Detail & Related papers (2023-11-06T08:52:11Z) - An In-Context Schema Understanding Method for Knowledge Base Question
Answering [70.87993081445127]
Large Language Models (LLMs) have shown strong capabilities in language understanding and can be used to solve this task.
Existing methods bypass this challenge by initially employing LLMs to generate drafts of logic forms without schema-specific details.
We propose a simple In-Context Understanding (ICSU) method that enables LLMs to directly understand schemas by leveraging in-context learning.
arXiv Detail & Related papers (2023-10-22T04:19:17Z) - Spider4SPARQL: A Complex Benchmark for Evaluating Knowledge Graph
Question Answering Systems [1.4732811715354452]
It has become increasingly important to provide realistic benchmarks for evaluating Knowledge Graph Question Answering systems.
Spider4SPARQL is a new SPARQL benchmark dataset featuring 9,693 previously existing manually generated NL questions and 4,721 unique, novel, and complex SPARQL queries.
We evaluate the system with state-of-the-art KGQA systems as well as LLMs, which achieve only up to 45% execution accuracy.
arXiv Detail & Related papers (2023-09-28T08:41:08Z) - Knowledge Base Question Answering by Case-based Reasoning over Subgraphs [81.22050011503933]
We show that our model answers queries requiring complex reasoning patterns more effectively than existing KG completion algorithms.
The proposed model outperforms or performs competitively with state-of-the-art models on several KBQA benchmarks.
arXiv Detail & Related papers (2022-02-22T01:34:35Z) - SPBERT: Pre-training BERT on SPARQL Queries for End-to-end Question
Answering over Knowledge Graphs [1.1775939485654976]
SPBERT is a Transformer-based language model pre-trained on massive SPARQL query logs.
We investigate how SPBERT and encoder-decoder architecture can be adapted for Knowledge-based QA corpora.
arXiv Detail & Related papers (2021-06-18T08:39:26Z) - Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks [133.93803565077337]
retrieval-augmented generation models combine pre-trained parametric and non-parametric memory for language generation.
We show that RAG models generate more specific, diverse and factual language than a state-of-the-art parametric-only seq2seq baseline.
arXiv Detail & Related papers (2020-05-22T21:34:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.