KMIR: A Benchmark for Evaluating Knowledge Memorization, Identification
and Reasoning Abilities of Language Models
- URL: http://arxiv.org/abs/2202.13529v1
- Date: Mon, 28 Feb 2022 03:52:57 GMT
- Title: KMIR: A Benchmark for Evaluating Knowledge Memorization, Identification
and Reasoning Abilities of Language Models
- Authors: Daniel Gao, Yantao Jia, Lei Li, Chengzhen Fu, Zhicheng Dou, Hao Jiang,
Xinyu Zhang, Lei Chen, Zhao Cao
- Abstract summary: We propose a benchmark, named Knowledge Memorization, Identification, and Reasoning test (KMIR)
KMIR covers 3 types of knowledge, including general knowledge, domain-specific knowledge, and commonsense, and provides 184,348 well-designed questions.
Preliminary experiments with various representative pre-training language models on KMIR reveal many interesting phenomenons.
- Score: 28.82149012250609
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Previous works show the great potential of pre-trained language models (PLMs)
for storing a large amount of factual knowledge. However, to figure out whether
PLMs can be reliable knowledge sources and used as alternative knowledge bases
(KBs), we need to further explore some critical features of PLMs. Firstly,
knowledge memorization and identification abilities: traditional KBs can store
various types of entities and relationships; do PLMs have a high knowledge
capacity to store different types of knowledge? Secondly, reasoning ability: a
qualified knowledge source should not only provide a collection of facts, but
support a symbolic reasoner. Can PLMs derive new knowledge based on the
correlations between facts? To evaluate these features of PLMs, we propose a
benchmark, named Knowledge Memorization, Identification, and Reasoning test
(KMIR). KMIR covers 3 types of knowledge, including general knowledge,
domain-specific knowledge, and commonsense, and provides 184,348 well-designed
questions. Preliminary experiments with various representative pre-training
language models on KMIR reveal many interesting phenomenons: 1) The
memorization ability of PLMs depends more on the number of parameters than
training schemes. 2) Current PLMs are struggling to robustly remember the
facts. 3) Model compression technology retains the amount of knowledge well,
but hurts the identification and reasoning abilities. We hope KMIR can
facilitate the design of PLMs as better knowledge sources.
Related papers
- What Matters in Memorizing and Recalling Facts? Multifaceted Benchmarks for Knowledge Probing in Language Models [15.057992220389604]
Language models often struggle with handling factual knowledge, exhibiting factual hallucination issue.
We introduce a knowledge probing benchmark, BELIEF(ICL), to evaluate the knowledge recall ability of both encoder- and decoder-based pre-trained language models.
We semi-automatically create MyriadLAMA, which has massively diverse prompts.
arXiv Detail & Related papers (2024-06-18T05:11:35Z) - Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction [15.534647327246239]
We propose an approach for estimating the latent knowledge embedded inside large language models (LLMs)
We leverage the in-context learning abilities of LLMs to estimate the extent to which an LLM knows the facts stored in a knowledge base.
arXiv Detail & Related papers (2024-04-19T15:40:39Z) - Can Language Models Act as Knowledge Bases at Scale? [24.99538360485476]
Large language models (LLMs) have demonstrated remarkable proficiency in understanding and generating responses to complex queries.
Our research investigates whether LLMs can effectively store, recall, and reason with knowledge on a large scale comparable to latest knowledge bases (KBs) such as Wikidata.
arXiv Detail & Related papers (2024-02-22T04:20:14Z) - Do PLMs Know and Understand Ontological Knowledge? [72.48752398867651]
Ontological knowledge comprises classes and properties and their relationships.
It is significant to explore whether Pretrained Language Models (PLMs) know and understand such knowledge.
Our results show that PLMs can memorize certain ontological knowledge and utilize implicit knowledge in reasoning.
arXiv Detail & Related papers (2023-09-12T03:20:50Z) - KnowledGPT: Enhancing Large Language Models with Retrieval and Storage
Access on Knowledge Bases [55.942342665806656]
KnowledGPT is a comprehensive framework to bridge large language models with various knowledge bases.
The retrieval process employs the program of thought prompting, which generates search language for KBs in code format.
KnowledGPT offers the capability to store knowledge in a personalized KB, catering to individual user demands.
arXiv Detail & Related papers (2023-08-17T13:07:00Z) - Do Large Language Models Know What They Don't Know? [74.65014158544011]
Large language models (LLMs) have a wealth of knowledge that allows them to excel in various Natural Language Processing (NLP) tasks.
Despite their vast knowledge, LLMs are still limited by the amount of information they can accommodate and comprehend.
This study aims to evaluate LLMs' self-knowledge by assessing their ability to identify unanswerable or unknowable questions.
arXiv Detail & Related papers (2023-05-29T15:30:13Z) - Measuring the Knowledge Acquisition-Utilization Gap in Pretrained
Language Models [26.342351417963965]
Pre-trained language models (PLMs) have shown evidence of acquiring vast amounts of knowledge.
It remains unclear how much of this parametric knowledge is actually usable in performing downstream tasks.
We propose a systematic framework to measure parametric knowledge utilization in PLMs.
arXiv Detail & Related papers (2023-05-24T06:26:11Z) - Decouple knowledge from parameters for plug-and-play language modeling [77.5601135412186]
We introduce PlugLM, a pre-training model with differentiable plug-in memory(DPM)
The key intuition is to decouple the knowledge storage from model parameters with an editable and scalable key-value memory.
PlugLM obtains 3.95 F1 improvements across four domains on average without any in-domain pre-training.
arXiv Detail & Related papers (2023-05-19T10:01:55Z) - Knowledge Rumination for Pre-trained Language Models [77.55888291165462]
We propose a new paradigm dubbed Knowledge Rumination to help the pre-trained language model utilize related latent knowledge without retrieving it from the external corpus.
We apply the proposed knowledge rumination to various language models, including RoBERTa, DeBERTa, and GPT-3.
arXiv Detail & Related papers (2023-05-15T15:47:09Z) - COPEN: Probing Conceptual Knowledge in Pre-trained Language Models [60.10147136876669]
Conceptual knowledge is fundamental to human cognition and knowledge bases.
Existing knowledge probing works only focus on factual knowledge of pre-trained language models (PLMs) and ignore conceptual knowledge.
We design three tasks to probe whether PLMs organize entities by conceptual similarities, learn conceptual properties, and conceptualize entities in contexts.
For the tasks, we collect and annotate 24k data instances covering 393 concepts, which is COPEN, a COnceptual knowledge Probing bENchmark.
arXiv Detail & Related papers (2022-11-08T08:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.