Relational world knowledge representation in contextual language models:
A review
- URL: http://arxiv.org/abs/2104.05837v1
- Date: Mon, 12 Apr 2021 21:50:55 GMT
- Title: Relational world knowledge representation in contextual language models:
A review
- Authors: Tara Safavi, Danai Koutra
- Abstract summary: We take a natural language processing perspective to the limitations of knowledge bases (KBs)
We propose a novel taxonomy for relational knowledge representation in contextual language models (LMs)
- Score: 19.176173014629185
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Relational knowledge bases (KBs) are established tools for world knowledge
representation in machines. While they are advantageous for their precision and
interpretability, they usually sacrifice some data modeling flexibility for
these advantages because they adhere to a manually engineered schema. In this
review, we take a natural language processing perspective to the limitations of
KBs, examining how they may be addressed in part by training neural contextual
language models (LMs) to internalize and express relational knowledge in
free-text form. We propose a novel taxonomy for relational knowledge
representation in contextual LMs based on the level of KB supervision provided,
considering both works that probe LMs for implicit relational knowledge
acquired during self-supervised pretraining on unstructured text alone, and
works that explicitly supervise LMs at the level of KB entities and/or
relations. We conclude that LMs and KBs are complementary representation tools,
as KBs provide a high standard of factual precision which can in turn be
flexibly and expressively modeled by LMs, and provide suggestions for future
research in this direction.
Related papers
- A Knowledge-Injected Curriculum Pretraining Framework for Question Answering [70.13026036388794]
We propose a general Knowledge-Injected Curriculum Pretraining framework (KICP) to achieve comprehensive KG learning and exploitation for Knowledge-based question answering tasks.
The KI module first injects knowledge into the LM by generating KG-centered pretraining corpus, and generalizes the process into three key steps.
The KA module learns knowledge from the generated corpus with LM equipped with an adapter as well as keeps its original natural language understanding ability.
The CR module follows human reasoning patterns to construct three corpora with increasing difficulties of reasoning, and further trains the LM from easy to hard in a curriculum manner.
arXiv Detail & Related papers (2024-03-11T03:42:03Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - Bridging the KB-Text Gap: Leveraging Structured Knowledge-aware
Pre-training for KBQA [28.642711264323786]
We propose a Structured Knowledge-aware Pre-training method (SKP) to bridge the gap between texts and structured KBs.
In the pre-training stage, we introduce two novel structured knowledge-aware tasks, guiding the model to effectively learn the implicit relationship and better representations of complex subgraphs.
In the downstream KBQA task, we further design an efficient linearization strategy and an interval attention mechanism, which assist the model to better encode complex subgraphs.
arXiv Detail & Related papers (2023-08-28T09:22:02Z) - Large Language Models with Controllable Working Memory [64.71038763708161]
Large language models (LLMs) have led to a series of breakthroughs in natural language processing (NLP)
What further sets these models apart is the massive amounts of world knowledge they internalize during pretraining.
How the model's world knowledge interacts with the factual information presented in the context remains under explored.
arXiv Detail & Related papers (2022-11-09T18:58:29Z) - Knowledge Prompting in Pre-trained Language Model for Natural Language
Understanding [24.315130086787374]
We propose a knowledge-prompting-based PLM framework KP-PLM.
This framework can be flexibly combined with existing mainstream PLMs.
To further leverage the factual knowledge from these prompts, we propose two novel knowledge-aware self-supervised tasks.
arXiv Detail & Related papers (2022-10-16T13:36:57Z) - BertNet: Harvesting Knowledge Graphs with Arbitrary Relations from
Pretrained Language Models [65.51390418485207]
We propose a new approach of harvesting massive KGs of arbitrary relations from pretrained LMs.
With minimal input of a relation definition, the approach efficiently searches in the vast entity pair space to extract diverse accurate knowledge.
We deploy the approach to harvest KGs of over 400 new relations from different LMs.
arXiv Detail & Related papers (2022-06-28T19:46:29Z) - Language Models as Knowledge Embeddings [26.384327693266837]
We propose LMKE, which adopts Language Models to derive Knowledge Embeddings.
We formulate description-based KE learning with a contrastive learning framework to improve efficiency in training and evaluation.
arXiv Detail & Related papers (2022-06-25T10:39:12Z) - A Review on Language Models as Knowledge Bases [55.035030134703995]
Recently, there has been a surge of interest in the NLP community on the use of pretrained Language Models (LMs) as Knowledge Bases (KBs)
arXiv Detail & Related papers (2022-04-12T18:35:23Z) - Reasoning over Vision and Language: Exploring the Benefits of
Supplemental Knowledge [59.87823082513752]
This paper investigates the injection of knowledge from general-purpose knowledge bases (KBs) into vision-and-language transformers.
We empirically study the relevance of various KBs to multiple tasks and benchmarks.
The technique is model-agnostic and can expand the applicability of any vision-and-language transformer with minimal computational overhead.
arXiv Detail & Related papers (2021-01-15T08:37:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.