KEPLET: Knowledge-Enhanced Pretrained Language Model with Topic Entity
Awareness
- URL: http://arxiv.org/abs/2305.01810v1
- Date: Tue, 2 May 2023 22:28:26 GMT
- Title: KEPLET: Knowledge-Enhanced Pretrained Language Model with Topic Entity
Awareness
- Authors: Yichuan Li, Jialong Han, Kyumin Lee, Chengyuan Ma, Benjamin Yao, Derek
Liu
- Abstract summary: We propose KEPLET, a Knowledge-Enhanced Pre-trained LanguagE model with Topic entity awareness.
In an end-to-end manner, KEPLET identifies where to add the topic entity's information in a Wikipedia sentence.
Experiments demonstrated the generality and superiority of KEPLET which was applied to two representative KEPLMs.
- Score: 12.90996504014071
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, Pre-trained Language Models (PLMs) have shown their
superiority by pre-training on unstructured text corpus and then fine-tuning on
downstream tasks. On entity-rich textual resources like Wikipedia,
Knowledge-Enhanced PLMs (KEPLMs) incorporate the interactions between tokens
and mentioned entities in pre-training, and are thus more effective on
entity-centric tasks such as entity linking and relation classification.
Although exploiting Wikipedia's rich structures to some extent, conventional
KEPLMs still neglect a unique layout of the corpus where each Wikipedia page is
around a topic entity (identified by the page URL and shown in the page title).
In this paper, we demonstrate that KEPLMs without incorporating the topic
entities will lead to insufficient entity interaction and biased (relation)
word semantics. We thus propose KEPLET, a novel Knowledge-Enhanced Pre-trained
LanguagE model with Topic entity awareness. In an end-to-end manner, KEPLET
identifies where to add the topic entity's information in a Wikipedia sentence,
fuses such information into token and mentioned entities representations, and
supervises the network learning, through which it takes topic entities back
into consideration. Experiments demonstrated the generality and superiority of
KEPLET which was applied to two representative KEPLMs, achieving significant
improvements on four entity-centric tasks.
Related papers
- Exploiting the Semantic Knowledge of Pre-trained Text-Encoders for Continual Learning [70.64617500380287]
Continual learning allows models to learn from new data while retaining previously learned knowledge.
The semantic knowledge available in the label information of the images, offers important semantic information that can be related with previously acquired knowledge of semantic classes.
We propose integrating semantic guidance within and across tasks by capturing semantic similarity using text embeddings.
arXiv Detail & Related papers (2024-08-02T07:51:44Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - Wikiformer: Pre-training with Structured Information of Wikipedia for
Ad-hoc Retrieval [21.262531222066208]
In this paper, we devise four pre-training objectives tailored for information retrieval tasks based on the structured knowledge of Wikipedia.
Compared to existing pre-training methods, our approach can better capture the semantic knowledge in the training corpus.
Experimental results in biomedical and legal domains demonstrate that our approach achieves better performance in vertical domains.
arXiv Detail & Related papers (2023-12-17T09:31:47Z) - Knowledge-Rich Self-Supervised Entity Linking [58.838404666183656]
Knowledge-RIch Self-Supervision ($tt KRISSBERT$) is a universal entity linker for four million UMLS entities.
Our approach subsumes zero-shot and few-shot methods, and can easily incorporate entity descriptions and gold mention labels if available.
Without using any labeled information, our method produces $tt KRISSBERT$, a universal entity linker for four million UMLS entities.
arXiv Detail & Related papers (2021-12-15T05:05:12Z) - Improving Entity Linking through Semantic Reinforced Entity Embeddings [16.868791358905916]
We propose a method to inject fine-grained semantic information into entity embeddings to reduce the distinctiveness and facilitate the learning of contextual commonality.
Based on our entity embeddings, we achieved new sate-of-the-art performance on entity linking.
arXiv Detail & Related papers (2021-06-16T00:27:56Z) - KI-BERT: Infusing Knowledge Context for Better Language and Domain
Understanding [0.0]
We propose a technique to infuse knowledge context from knowledge graphs for conceptual and ambiguous entities into models based on transformer architecture.
Our novel technique project knowledge graph embedding in the homogeneous vector-space, introduces new token-types for entities, align entity position ids, and a selective attention mechanism.
We take BERT as a baseline model and implement "KnowledgeInfused BERT" by infusing knowledge context from ConceptNet and WordNet.
arXiv Detail & Related papers (2021-04-09T16:15:31Z) - ERICA: Improving Entity and Relation Understanding for Pre-trained
Language Models via Contrastive Learning [97.10875695679499]
We propose a novel contrastive learning framework named ERICA in pre-training phase to obtain a deeper understanding of the entities and their relations in text.
Experimental results demonstrate that our proposed ERICA framework achieves consistent improvements on several document-level language understanding tasks.
arXiv Detail & Related papers (2020-12-30T03:35:22Z) - LUKE: Deep Contextualized Entity Representations with Entity-aware
Self-attention [37.111204321059084]
We propose new pretrained contextualized representations of words and entities based on the bidirectional transformer.
Our model is trained using a new pretraining task based on the masked language model of BERT.
We also propose an entity-aware self-attention mechanism that is an extension of the self-attention mechanism of the transformer.
arXiv Detail & Related papers (2020-10-02T15:38:03Z) - Autoregressive Entity Retrieval [55.38027440347138]
Entities are at the center of how we represent and aggregate knowledge.
The ability to retrieve such entities given a query is fundamental for knowledge-intensive tasks such as entity linking and open-domain question answering.
We propose GENRE, the first system that retrieves entities by generating their unique names, left to right, token-by-token in an autoregressive fashion.
arXiv Detail & Related papers (2020-10-02T10:13:31Z) - CoLAKE: Contextualized Language and Knowledge Embedding [81.90416952762803]
We propose the Contextualized Language and Knowledge Embedding (CoLAKE)
CoLAKE jointly learns contextualized representation for both language and knowledge with the extended objective.
We conduct experiments on knowledge-driven tasks, knowledge probing tasks, and language understanding tasks.
arXiv Detail & Related papers (2020-10-01T11:39:32Z) - Learning Cross-Context Entity Representations from Text [9.981223356176496]
We investigate the use of a fill-in-the-blank task to learn context independent representations of entities from text contexts.
We show that large scale training of neural models allows us to learn high quality entity representations.
Our global entity representations encode fine-grained type categories, such as Scottish footballers, and can answer trivia questions.
arXiv Detail & Related papers (2020-01-11T15:30:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.