Improving Entity Disambiguation by Reasoning over a Knowledge Base
- URL: http://arxiv.org/abs/2207.04106v1
- Date: Fri, 8 Jul 2022 19:13:53 GMT
- Title: Improving Entity Disambiguation by Reasoning over a Knowledge Base
- Authors: Tom Ayoola, Joseph Fisher, Andrea Pierleoni
- Abstract summary: We introduce an ED model which links entities by reasoning over a symbolic knowledge base.
Our model surpasses state-of-the-art baselines on six well-established ED datasets by 1.3 F1 on average.
- Score: 2.223733768286313
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work in entity disambiguation (ED) has typically neglected structured
knowledge base (KB) facts, and instead relied on a limited subset of KB
information, such as entity descriptions or types. This limits the range of
contexts in which entities can be disambiguated. To allow the use of all KB
facts, as well as descriptions and types, we introduce an ED model which links
entities by reasoning over a symbolic knowledge base in a fully differentiable
fashion. Our model surpasses state-of-the-art baselines on six well-established
ED datasets by 1.3 F1 on average. By allowing access to all KB information, our
model is less reliant on popularity-based entity priors, and improves
performance on the challenging ShadowLink dataset (which emphasises infrequent
and ambiguous entities) by 12.7 F1.
Related papers
- OneNet: A Fine-Tuning Free Framework for Few-Shot Entity Linking via Large Language Model Prompting [49.655711022673046]
OneNet is an innovative framework that utilizes the few-shot learning capabilities of Large Language Models (LLMs) without the need for fine-tuning.
OneNet is structured around three key components prompted by LLMs: (1) an entity reduction processor that simplifies inputs by summarizing and filtering out irrelevant entities, (2) a dual-perspective entity linker that combines contextual cues and prior knowledge for precise entity linking, and (3) an entity consensus judger that employs a unique consistency algorithm to alleviate the hallucination in the entity linking reasoning.
arXiv Detail & Related papers (2024-10-10T02:45:23Z) - Real World Conversational Entity Linking Requires More Than Zeroshots [50.5691094768954]
We design targeted evaluation scenarios to measure the efficacy of EL models under resource constraints.
We assess EL models' ability to generalize to a new unfamiliar KB using Fandom and a novel zero-shot conversational entity linking dataset.
Results indicate that current zero-shot EL models falter when introduced to new, domain-specific KBs without prior training.
arXiv Detail & Related papers (2024-09-02T10:37:53Z) - Entity Disambiguation via Fusion Entity Decoding [68.77265315142296]
We propose an encoder-decoder model to disambiguate entities with more detailed entity descriptions.
We observe +1.5% improvements in end-to-end entity linking in the GERBIL benchmark compared with EntQA.
arXiv Detail & Related papers (2024-04-02T04:27:54Z) - MoCoSA: Momentum Contrast for Knowledge Graph Completion with
Structure-Augmented Pre-trained Language Models [11.57782182864771]
We propose Momentum Contrast for knowledge graph completion with Structure-Augmented pre-trained language models (MoCoSA)
Our approach achieves state-of-the-art performance in terms of mean reciprocal rank (MRR), with improvements of 2.5% on WN18RR and 21% on OpenBG500.
arXiv Detail & Related papers (2023-08-16T08:09:10Z) - FactKB: Generalizable Factuality Evaluation using Language Models
Enhanced with Factual Knowledge [37.2179237007464]
We propose FactKB, a simple new approach to factuality evaluation that is generalizable across domains.
We introduce three types of complementary factuality pretraining objectives based on direct entity facts, facts grounded in auxiliary knowledge about entities, and facts constructed compositionally through knowledge base walks.
The resulting factuality evaluation model achieves state-of-the-art performance on two in-domain news summarization benchmarks and on three out-of-domain scientific literature datasets.
arXiv Detail & Related papers (2023-05-14T23:58:05Z) - Modeling Entities as Semantic Points for Visual Information Extraction
in the Wild [55.91783742370978]
We propose an alternative approach to precisely and robustly extract key information from document images.
We explicitly model entities as semantic points, i.e., center points of entities are enriched with semantic information describing the attributes and relationships of different entities.
The proposed method can achieve significantly enhanced performance on entity labeling and linking, compared with previous state-of-the-art models.
arXiv Detail & Related papers (2023-03-23T08:21:16Z) - Named Entity Linking on Namesakes [10.609815608017065]
We represent knowledge base (KB) entity by a set of embeddings.
We show that representations of entities in the knowledge base (KB) can be adjusted using only KB data, and the adjustment improves NEL performance.
arXiv Detail & Related papers (2022-05-21T03:31:25Z) - Knowledge-Rich Self-Supervised Entity Linking [58.838404666183656]
Knowledge-RIch Self-Supervision ($tt KRISSBERT$) is a universal entity linker for four million UMLS entities.
Our approach subsumes zero-shot and few-shot methods, and can easily incorporate entity descriptions and gold mention labels if available.
Without using any labeled information, our method produces $tt KRISSBERT$, a universal entity linker for four million UMLS entities.
arXiv Detail & Related papers (2021-12-15T05:05:12Z) - Metadata Shaping: Natural Language Annotations for the Tail [4.665656172490747]
Language models (LMs) have made remarkable progress, but still struggle to generalize beyond the training data to rare linguistic patterns.
We propose metadata shaping, a method in which readily available metadata, such as entity descriptions and categorical tags, are appended to examples based on information theoretic metrics.
With no changes to the LM whatsoever, metadata shaping exceeds the BERT-baseline by up to 5.3 F1 points, and achieves or competes with state-of-the-art results.
arXiv Detail & Related papers (2021-10-16T01:00:47Z) - Linking Entities to Unseen Knowledge Bases with Arbitrary Schemas [31.154104663488358]
In entity linking, mentions of named entities in raw text are disambiguated against a knowledge base (KB)
This work focuses on linking to unseen KBs that do not have training data and whose schema is unknown during training.
Our approach relies on methods to flexibly convert entities from arbitrary KBs with several attribute-value pairs into flat strings.
arXiv Detail & Related papers (2020-10-21T22:07:31Z) - BoxE: A Box Embedding Model for Knowledge Base Completion [53.57588201197374]
Knowledge base completion (KBC) aims to automatically infer missing facts by exploiting information already present in a knowledge base (KB)
Existing embedding models are subject to at least one of the following limitations.
BoxE embeds entities as points, and relations as a set of hyper-rectangles (or boxes)
arXiv Detail & Related papers (2020-07-13T09:40:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.