Knowledge Graph Question Answering Datasets and Their Generalizability:
Are They Enough for Future Research?
- URL: http://arxiv.org/abs/2205.06573v1
- Date: Fri, 13 May 2022 12:01:15 GMT
- Title: Knowledge Graph Question Answering Datasets and Their Generalizability:
Are They Enough for Future Research?
- Authors: Longquan Jiang, Ricardo Usbeck
- Abstract summary: We analyze 25 well-known KGQA datasets for 5 different Knowledge Graphs (KGs)
We show that according to this definition many existing and online available KGQA datasets are either not suited to train a generalizable KGQA system or that the datasets are based on discontinued and out-dated KGs.
We propose a mitigation method for re-splitting available KGQA datasets to enable their applicability to evaluate generalization, without any cost and manual effort.
- Score: 0.7817685358710509
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing approaches on Question Answering over Knowledge Graphs (KGQA) have
weak generalizability. That is often due to the standard i.i.d. assumption on
the underlying dataset. Recently, three levels of generalization for KGQA were
defined, namely i.i.d., compositional, zero-shot. We analyze 25 well-known KGQA
datasets for 5 different Knowledge Graphs (KGs). We show that according to this
definition many existing and online available KGQA datasets are either not
suited to train a generalizable KGQA system or that the datasets are based on
discontinued and out-dated KGs. Generating new datasets is a costly process
and, thus, is not an alternative to smaller research groups and companies. In
this work, we propose a mitigation method for re-splitting available KGQA
datasets to enable their applicability to evaluate generalization, without any
cost and manual effort. We test our hypothesis on three KGQA datasets, i.e.,
LC-QuAD, LC-QuAD 2.0 and QALD-9). Experiments on re-splitted KGQA datasets
demonstrate its effectiveness towards generalizability. The code and a unified
way to access 18 available datasets is online at
https://github.com/semantic-systems/KGQA-datasets as well as
https://github.com/semantic-systems/KGQA-datasets-generalization.
Related papers
- Generate-on-Graph: Treat LLM as both Agent and KG in Incomplete Knowledge Graph Question Answering [87.67177556994525]
We propose a training-free method called Generate-on-Graph (GoG) to generate new factual triples while exploring Knowledge Graphs (KGs)
GoG performs reasoning through a Thinking-Searching-Generating framework, which treats LLM as both Agent and KG in IKGQA.
arXiv Detail & Related papers (2024-04-23T04:47:22Z) - CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring
Commonsense Reasoning and Long-Tail Knowledge [21.73770363188049]
We create a novel Commonsense Reasoning (CR) and Long-Tail (LT) KGQA dataset with two subtasks -- question answering and claim verification.
While existing KGQA methods are not applicable due to their lack of commonsense inference support, baseline evaluation of LLMs on CR-LT KGQA demonstrate a high rate of hallucination.
arXiv Detail & Related papers (2024-03-03T04:47:01Z) - GSQA: An End-to-End Model for Generative Spoken Question Answering [54.418723701886115]
We introduce the first end-to-end Generative Spoken Question Answering (GSQA) model that empowers the system to engage in abstractive reasoning.
Our model surpasses the previous extractive model by 3% on extractive QA datasets.
Our GSQA model shows the potential to generalize to a broad spectrum of questions, thus further expanding the spoken question answering capabilities of abstractive QA.
arXiv Detail & Related papers (2023-12-15T13:33:18Z) - PyGraft: Configurable Generation of Synthetic Schemas and Knowledge
Graphs at Your Fingertips [3.5923669681271257]
PyGraft is a Python-based tool that generates customized, domain-agnostic schemas and KGs.
We aim to empower the generation of a more diverse array of KGs for benchmarking novel approaches in areas such as graph-based machine learning (ML)
In ML, this should foster a more holistic evaluation of model performance and generalization capability, thereby going beyond the limited collection of available benchmarks.
arXiv Detail & Related papers (2023-09-07T13:00:09Z) - KGxBoard: Explainable and Interactive Leaderboard for Evaluation of
Knowledge Graph Completion Models [76.01814380927507]
KGxBoard is an interactive framework for performing fine-grained evaluation on meaningful subsets of the data.
In our experiments, we highlight the findings with the use of KGxBoard, which would have been impossible to detect with standard averaged single-score metrics.
arXiv Detail & Related papers (2022-08-23T15:11:45Z) - Knowledge Graph Question Answering Leaderboard: A Community Resource to
Prevent a Replication Crisis [61.740077541531726]
We provide a new central and open leaderboard for any KGQA benchmark dataset as a focal point for the community.
Our analysis highlights existing problems during the evaluation of KGQA systems.
arXiv Detail & Related papers (2022-01-20T13:46:01Z) - HeteroQA: Learning towards Question-and-Answering through Multiple
Information Sources via Heterogeneous Graph Modeling [50.39787601462344]
Community Question Answering (CQA) is a well-defined task that can be used in many scenarios, such as E-Commerce and online user community for special interests.
Most of the CQA methods only incorporate articles or Wikipedia to extract knowledge and answer the user's question.
We propose a question-aware heterogeneous graph transformer to incorporate the multiple information sources (MIS) in the user community to automatically generate the answer.
arXiv Detail & Related papers (2021-12-27T10:16:43Z) - Question Answering Over Temporal Knowledge Graphs [20.479222151497495]
Temporal Knowledge Graphs (Temporal KGs) extend regular Knowledge Graphs by providing temporal scopes (start and end times) on each edge in the KG.
While Question Answering over KG (KGQA) has received some attention from the research community, QA over Temporal KGs (Temporal KGQA) is a relatively unexplored area.
We present CRONQUESTIONS, the largest known Temporal KGQA dataset, clearly stratified into buckets of structural complexity.
arXiv Detail & Related papers (2021-06-03T00:45:07Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z) - What do Models Learn from Question Answering Datasets? [2.28438857884398]
We investigate if models are learning reading comprehension from question answering datasets.
We evaluate models on their generalizability to out-of-domain examples, responses to missing or incorrect data, and ability to handle question variations.
We make recommendations for building future QA datasets that better evaluate the task of question answering through reading comprehension.
arXiv Detail & Related papers (2020-04-07T15:41:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.