Neural Entity Linking: A Survey of Models Based on Deep Learning
- URL: http://arxiv.org/abs/2006.00575v4
- Date: Thu, 7 Apr 2022 17:56:30 GMT
- Title: Neural Entity Linking: A Survey of Models Based on Deep Learning
- Authors: Ozge Sevgili, Artem Shelmanov, Mikhail Arkhipov, Alexander Panchenko,
Chris Biemann
- Abstract summary: This survey presents a comprehensive description of recent neural entity linking (EL) systems developed since 2015.
Its goal is to systemize design features of neural entity linking systems and compare their performance to the remarkable classic methods on common benchmarks.
The survey touches on applications of entity linking, focusing on the recently emerged use-case of enhancing deep pre-trained masked language models.
- Score: 82.43751915717225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This survey presents a comprehensive description of recent neural entity
linking (EL) systems developed since 2015 as a result of the "deep learning
revolution" in natural language processing. Its goal is to systemize design
features of neural entity linking systems and compare their performance to the
remarkable classic methods on common benchmarks. This work distills a generic
architecture of a neural EL system and discusses its components, such as
candidate generation, mention-context encoding, and entity ranking, summarizing
prominent methods for each of them. The vast variety of modifications of this
general architecture are grouped by several common themes: joint entity mention
detection and disambiguation, models for global linking, domain-independent
techniques including zero-shot and distant supervision methods, and
cross-lingual approaches. Since many neural models take advantage of entity and
mention/context embeddings to represent their meaning, this work also overviews
prominent entity embedding techniques. Finally, the survey touches on
applications of entity linking, focusing on the recently emerged use-case of
enhancing deep pre-trained masked language models based on the Transformer
architecture.
Related papers
- A VAE-based Framework for Learning Multi-Level Neural Granger-Causal
Connectivity [15.295157876811066]
This paper introduces a Variational Autoencoder based framework that jointly learns Granger-causal relationships amongst components in a collection of related-yet-heterogeneous dynamical systems.
The performance of the proposed framework is evaluated on several synthetic data settings and benchmarked against existing approaches designed for individual system learning.
arXiv Detail & Related papers (2024-02-25T16:11:32Z) - Foundational Models Defining a New Era in Vision: A Survey and Outlook [151.49434496615427]
Vision systems to see and reason about the compositional nature of visual scenes are fundamental to understanding our world.
The models learned to bridge the gap between such modalities coupled with large-scale training data facilitate contextual reasoning, generalization, and prompt capabilities at test time.
The output of such models can be modified through human-provided prompts without retraining, e.g., segmenting a particular object by providing a bounding box, having interactive dialogues by asking questions about an image or video scene or manipulating the robot's behavior through language instructions.
arXiv Detail & Related papers (2023-07-25T17:59:18Z) - Neural Constraint Satisfaction: Hierarchical Abstraction for
Combinatorial Generalization in Object Rearrangement [75.9289887536165]
We present a hierarchical abstraction approach to uncover underlying entities.
We show how to learn a correspondence between intervening on states of entities in the agent's model and acting on objects in the environment.
We use this correspondence to develop a method for control that generalizes to different numbers and configurations of objects.
arXiv Detail & Related papers (2023-03-20T18:19:36Z) - Topics as Entity Clusters: Entity-based Topics from Large Language Models and Graph Neural Networks [0.6486052012623045]
We propose a novel topic clustering approach using bimodal vector representations of entities.
Our approach is better suited to working with entities in comparison to state-of-the-art models.
arXiv Detail & Related papers (2023-01-06T10:54:54Z) - A Review of Sparse Expert Models in Deep Learning [23.721204843236006]
Sparse expert models are a thirty-year old concept re-emerging as a popular architecture in deep learning.
We review the concept of sparse expert models, provide a basic description of the common algorithms, and contextualize the advances in the deep learning era.
arXiv Detail & Related papers (2022-09-04T18:00:29Z) - Redefining Neural Architecture Search of Heterogeneous Multi-Network
Models by Characterizing Variation Operators and Model Components [71.03032589756434]
We investigate the effect of different variation operators in a complex domain, that of multi-network heterogeneous neural models.
We characterize both the variation operators, according to their effect on the complexity and performance of the model; and the models, relying on diverse metrics which estimate the quality of the different parts composing it.
arXiv Detail & Related papers (2021-06-16T17:12:26Z) - Polynomial Networks in Deep Classifiers [55.90321402256631]
We cast the study of deep neural networks under a unifying framework.
Our framework provides insights on the inductive biases of each model.
The efficacy of the proposed models is evaluated on standard image and audio classification benchmarks.
arXiv Detail & Related papers (2021-04-16T06:41:20Z) - Model-Based Machine Learning for Communications [110.47840878388453]
We review existing strategies for combining model-based algorithms and machine learning from a high level perspective.
We focus on symbol detection, which is one of the fundamental tasks of communication receivers.
arXiv Detail & Related papers (2021-01-12T19:55:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.