Better Together -- An Ensemble Learner for Combining the Results of
Ready-made Entity Linking Systems
- URL: http://arxiv.org/abs/2101.05634v1
- Date: Thu, 14 Jan 2021 14:42:57 GMT
- Title: Better Together -- An Ensemble Learner for Combining the Results of
Ready-made Entity Linking Systems
- Authors: Renato Stoffalette Jo\~ao and Pavlos Fafalios and Stefan Dietze
- Abstract summary: We argue that performance may be optimised by exploiting results from distinct EL systems on the same corpus.
In this paper, we introduce a supervised approach which exploits the output of multiple ready-made EL systems by predicting the correct link on a per-mention basis.
- Score: 2.163881720692685
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Entity linking (EL) is the task of automatically identifying entity mentions
in text and resolving them to a corresponding entity in a reference knowledge
base like Wikipedia. Throughout the past decade, a plethora of EL systems and
pipelines have become available, where performance of individual systems varies
heavily across corpora, languages or domains. Linking performance varies even
between different mentions in the same text corpus, where, for instance, some
EL approaches are better able to deal with short surface forms while others may
perform better when more context information is available. To this end, we
argue that performance may be optimised by exploiting results from distinct EL
systems on the same corpus, thereby leveraging their individual strengths on a
per-mention basis. In this paper, we introduce a supervised approach which
exploits the output of multiple ready-made EL systems by predicting the correct
link on a per-mention basis. Experimental results obtained on existing ground
truth datasets and exploiting three state-of-the-art EL systems show the
effectiveness of our approach and its capacity to significantly outperform the
individual EL systems as well as a set of baseline methods.
Related papers
- Self-MoE: Towards Compositional Large Language Models with Self-Specialized Experts [49.950419707905944]
We present Self-MoE, an approach that transforms a monolithic LLM into a compositional, modular system of self-specialized experts.
Our approach leverages self-specialization, which constructs expert modules using self-generated synthetic data.
Our findings highlight the critical role of modularity, the applicability of Self-MoE to multiple base LLMs, and the potential of self-improvement in achieving efficient, scalable, and adaptable systems.
arXiv Detail & Related papers (2024-06-17T19:06:54Z) - Multilingual Entity Linking Using Dense Retrieval [0.0]
In this thesis, we develop systems that are fast to train and operate in multiple languages.
Our work shows that building competitive neural network based EL systems that operate in multiple languages is possible even with limited resources.
arXiv Detail & Related papers (2024-05-13T18:57:27Z) - Self-Retrieval: End-to-End Information Retrieval with One Large Language Model [97.71181484082663]
We introduce Self-Retrieval, a novel end-to-end LLM-driven information retrieval architecture.
Self-Retrieval internalizes the retrieval corpus through self-supervised learning, transforms the retrieval process into sequential passage generation, and performs relevance assessment for reranking.
arXiv Detail & Related papers (2024-02-23T18:45:35Z) - Instructed Language Models with Retrievers Are Powerful Entity Linkers [87.16283281290053]
Instructed Generative Entity Linker (INSGENEL) is the first approach that enables casual language models to perform entity linking over knowledge bases.
INSGENEL outperforms previous generative alternatives with +6.8 F1 points gain on average.
arXiv Detail & Related papers (2023-11-06T16:38:51Z) - Compositional Exemplars for In-context Learning [21.961094715261133]
Large pretrained language models (LMs) have shown impressive In-Context Learning (ICL) ability.
We propose CEIL (Compositional Exemplars for In-context Learning) to model the interaction between the given input and in-context examples.
We validate CEIL on 12 classification and generation datasets from 7 distinct NLP tasks, including sentiment analysis, paraphrase detection, natural language inference, commonsense reasoning, open-domain question answering, code generation, and semantic parsing.
arXiv Detail & Related papers (2023-02-11T14:02:08Z) - Hybrid Rule-Neural Coreference Resolution System based on Actor-Critic
Learning [53.73316523766183]
Coreference resolution systems need to tackle two main tasks.
One task is to detect all of the potential mentions, the other is to learn the linking of an antecedent for each possible mention.
We propose a hybrid rule-neural coreference resolution system based on actor-critic learning.
arXiv Detail & Related papers (2022-12-20T08:55:47Z) - Unified Structure Generation for Universal Information Extraction [58.89057387608414]
UIE can universally model different IE tasks, adaptively generate targeted structures, and collaboratively learn general IE abilities from different knowledge sources.
Experiments show that UIE achieved the state-of-the-art performance on 4 IE tasks, 13 datasets, and on all supervised, low-resource, and few-shot settings.
arXiv Detail & Related papers (2022-03-23T08:49:29Z) - Entity Linking Meets Deep Learning: Techniques and Solutions [49.017379833990155]
We present a comprehensive review and analysis of existing deep learning based EL methods.
We propose a new taxonomy, which organizes existing DL based EL methods using three axes: embedding, feature, and algorithm.
We give a quantitative performance analysis of DL based EL methods over data sets.
arXiv Detail & Related papers (2021-09-26T07:57:38Z) - Robustness Evaluation of Entity Disambiguation Using Prior Probes:the
Case of Entity Overshadowing [11.513083693564466]
We evaluate and report the performance of popular entity linking systems on the ShadowLink benchmark.
Results show a considerable difference in accuracy between more and less common entities for all of the EL systems under evaluation.
arXiv Detail & Related papers (2021-08-24T20:54:56Z) - Interpretable and Low-Resource Entity Matching via Decoupling Feature
Learning from Decision Making [22.755892575582788]
Entity Matching aims at recognizing entity records that denote the same real-world object.
We propose a novel EM framework that consists of Heterogeneous Information Fusion (HIF) and Key Attribute Tree (KAT) Induction.
Our method is highly efficient and outperforms SOTA EM models in most cases.
arXiv Detail & Related papers (2021-06-08T08:27:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.