Entity Disambiguation with Entity Definitions
- URL: http://arxiv.org/abs/2210.05648v1
- Date: Tue, 11 Oct 2022 17:46:28 GMT
- Title: Entity Disambiguation with Entity Definitions
- Authors: Luigi Procopio, Simone Conia, Edoardo Barba, Roberto Navigli
- Abstract summary: Local models have recently attained astounding performances in Entity Disambiguation (ED)
Previous works limited their studies to using, as the textual representation of each candidate, only its Wikipedia title.
In this paper, we address this limitation and investigate to what extent more expressive textual representations can mitigate it.
We report a new state of the art on 2 out of 6 benchmarks we consider and strongly improve the generalization capability over unseen patterns.
- Score: 50.01142092276296
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Local models have recently attained astounding performances in Entity
Disambiguation (ED), with generative and extractive formulations being the most
promising research directions. However, previous works limited their studies to
using, as the textual representation of each candidate, only its Wikipedia
title. Although certainly effective, this strategy presents a few critical
issues, especially when titles are not sufficiently informative or
distinguishable from one another. In this paper, we address this limitation and
investigate to what extent more expressive textual representations can mitigate
it. We thoroughly evaluate our approach against standard benchmarks in ED and
find extractive formulations to be particularly well-suited to these
representations: we report a new state of the art on 2 out of 6 benchmarks we
consider and strongly improve the generalization capability over unseen
patterns. We release our code, data and model checkpoints at
https://github.com/SapienzaNLP/extend.
Related papers
- Beyond Coarse-Grained Matching in Video-Text Retrieval [50.799697216533914]
We introduce a new approach for fine-grained evaluation.
Our approach can be applied to existing datasets by automatically generating hard negative test captions.
Experiments on our fine-grained evaluations demonstrate that this approach enhances a model's ability to understand fine-grained differences.
arXiv Detail & Related papers (2024-10-16T09:42:29Z) - Coherent Entity Disambiguation via Modeling Topic and Categorical
Dependency [87.16283281290053]
Previous entity disambiguation (ED) methods adopt a discriminative paradigm, where prediction is made based on matching scores between mention context and candidate entities.
We propose CoherentED, an ED system equipped with novel designs aimed at enhancing the coherence of entity predictions.
We achieve new state-of-the-art results on popular ED benchmarks, with an average improvement of 1.3 F1 points.
arXiv Detail & Related papers (2023-11-06T16:40:13Z) - Revisiting Sparse Retrieval for Few-shot Entity Linking [33.15662306409253]
We propose an ELECTRA-based keyword extractor to denoise the mention context and construct a better query expression.
For training the extractor, we propose a distant supervision method to automatically generate training data based on overlapping tokens between mention contexts and entity descriptions.
Experimental results on the ZESHEL dataset demonstrate that the proposed method outperforms state-of-the-art models by a significant margin across all test domains.
arXiv Detail & Related papers (2023-10-19T03:51:10Z) - On Context Utilization in Summarization with Large Language Models [83.84459732796302]
Large language models (LLMs) excel in abstractive summarization tasks, delivering fluent and pertinent summaries.
Recent advancements have extended their capabilities to handle long-input contexts, exceeding 100k tokens.
We conduct the first comprehensive study on context utilization and position bias in summarization.
arXiv Detail & Related papers (2023-10-16T16:45:12Z) - Enriching Relation Extraction with OpenIE [70.52564277675056]
Relation extraction (RE) is a sub-discipline of information extraction (IE)
In this work, we explore how recent approaches for open information extraction (OpenIE) may help to improve the task of RE.
Our experiments over two annotated corpora, KnowledgeNet and FewRel, demonstrate the improved accuracy of our enriched models.
arXiv Detail & Related papers (2022-12-19T11:26:23Z) - Robust Candidate Generation for Entity Linking on Short Social Media
Texts [1.5006258585503875]
We show that in the domain of Tweets, such methods suffer as users often include informal spelling, limited context, and lack of specificity.
We demonstrate a hybrid solution using long contextual representation from Wikipedia, achieving 0.93 recall.
arXiv Detail & Related papers (2022-10-14T02:47:31Z) - A Closer Look at Debiased Temporal Sentence Grounding in Videos:
Dataset, Metric, and Approach [53.727460222955266]
Temporal Sentence Grounding in Videos (TSGV) aims to ground a natural language sentence in an untrimmed video.
Recent studies have found that current benchmark datasets may have obvious moment annotation biases.
We introduce a new evaluation metric "dR@n,IoU@m" that discounts the basic recall scores to alleviate the inflating evaluation caused by biased datasets.
arXiv Detail & Related papers (2022-03-10T08:58:18Z) - Partially Supervised Named Entity Recognition via the Expected Entity
Ratio Loss [2.9176780332337486]
We study learning named entity recognizers in the presence of missing entity annotations.
We propose a novel loss, the Expected Entity Ratio, to learn models in the presence of systematically missing tags.
We show that our approach is both theoretically sound and empirically useful.
arXiv Detail & Related papers (2021-08-16T16:53:39Z) - Analysis and Evaluation of Language Models for Word Sense Disambiguation [18.001457030065712]
Transformer-based language models have taken many fields in NLP by storm.
BERT can accurately capture high-level sense distinctions, even when a limited number of examples is available for each word sense.
BERT and its derivatives dominate most of the existing evaluation benchmarks.
arXiv Detail & Related papers (2020-08-26T15:07:07Z) - Extending Text Informativeness Measures to Passage Interestingness
Evaluation (Language Model vs. Word Embedding) [1.2998637003026272]
This paper defines the concept of Interestingness as a generalization of Informativeness.
We then study the ability of state of the art Informativeness measures to cope with this generalization.
We prove that the CLEF-INEX Tweet Contextualization 2012 Logarithm Similarity measure provides best results.
arXiv Detail & Related papers (2020-04-14T18:22:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.