MuVER: Improving First-Stage Entity Retrieval with Multi-View Entity
Representations
- URL: http://arxiv.org/abs/2109.05716v1
- Date: Mon, 13 Sep 2021 05:51:45 GMT
- Title: MuVER: Improving First-Stage Entity Retrieval with Multi-View Entity
Representations
- Authors: Xinyin Ma, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei
Huang, Weiming Lu
- Abstract summary: We propose a novel approach for entity retrieval that constructs multi-view representations for entity descriptions and approximates the optimal view for mentions via a searching method.
Our method achieves the state-of-the-art performance on ZESHEL and improves the quality of candidates on three standard Entity Linking datasets.
- Score: 28.28940043641958
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Entity retrieval, which aims at disambiguating mentions to canonical entities
from massive KBs, is essential for many tasks in natural language processing.
Recent progress in entity retrieval shows that the dual-encoder structure is a
powerful and efficient framework to nominate candidates if entities are only
identified by descriptions. However, they ignore the property that meanings of
entity mentions diverge in different contexts and are related to various
portions of descriptions, which are treated equally in previous works. In this
work, we propose Multi-View Entity Representations (MuVER), a novel approach
for entity retrieval that constructs multi-view representations for entity
descriptions and approximates the optimal view for mentions via a heuristic
searching method. Our method achieves the state-of-the-art performance on
ZESHEL and improves the quality of candidates on three standard Entity Linking
datasets
Related papers
- OneNet: A Fine-Tuning Free Framework for Few-Shot Entity Linking via Large Language Model Prompting [49.655711022673046]
OneNet is an innovative framework that utilizes the few-shot learning capabilities of Large Language Models (LLMs) without the need for fine-tuning.
OneNet is structured around three key components prompted by LLMs: (1) an entity reduction processor that simplifies inputs by summarizing and filtering out irrelevant entities, (2) a dual-perspective entity linker that combines contextual cues and prior knowledge for precise entity linking, and (3) an entity consensus judger that employs a unique consistency algorithm to alleviate the hallucination in the entity linking reasoning.
arXiv Detail & Related papers (2024-10-10T02:45:23Z) - Entity Disambiguation via Fusion Entity Decoding [68.77265315142296]
We propose an encoder-decoder model to disambiguate entities with more detailed entity descriptions.
We observe +1.5% improvements in end-to-end entity linking in the GERBIL benchmark compared with EntQA.
arXiv Detail & Related papers (2024-04-02T04:27:54Z) - LLMs as Bridges: Reformulating Grounded Multimodal Named Entity Recognition [28.136662420053568]
Grounded Multimodal Named Entity Recognition (GMNER) is a nascent multimodal task that aims to identify named entities, entity types and their corresponding visual regions.
We propose RiVEG, a unified framework that reformulates GMNER into a joint MNER-VE-VG task by leveraging large language models (LLMs) as a connecting bridge.
arXiv Detail & Related papers (2024-02-15T14:54:33Z) - Seed-Guided Fine-Grained Entity Typing in Science and Engineering
Domains [51.02035914828596]
We study the task of seed-guided fine-grained entity typing in science and engineering domains.
We propose SEType which first enriches the weak supervision by finding more entities for each seen type from an unlabeled corpus.
It then matches the enriched entities to unlabeled text to get pseudo-labeled samples and trains a textual entailment model that can make inferences for both seen and unseen types.
arXiv Detail & Related papers (2024-01-23T22:36:03Z) - Few-Shot Nested Named Entity Recognition [4.8693196802491405]
This paper is the first one dedicated to studying the few-shot nested NER task.
We propose a Biaffine-based Contrastive Learning (BCL) framework to learn contextual dependency to distinguish nested entities.
The BCL outperformed three baseline models on the 1-shot and 5-shot tasks in terms of F1 score.
arXiv Detail & Related papers (2022-12-02T03:42:23Z) - UnifieR: A Unified Retriever for Large-Scale Retrieval [84.61239936314597]
Large-scale retrieval is to recall relevant documents from a huge collection given a query.
Recent retrieval methods based on pre-trained language models (PLM) can be coarsely categorized into either dense-vector or lexicon-based paradigms.
We propose a new learning framework, UnifieR which unifies dense-vector and lexicon-based retrieval in one model with a dual-representing capability.
arXiv Detail & Related papers (2022-05-23T11:01:59Z) - Good Visual Guidance Makes A Better Extractor: Hierarchical Visual
Prefix for Multimodal Entity and Relation Extraction [88.6585431949086]
We propose a novel Hierarchical Visual Prefix fusion NeTwork (HVPNeT) for visual-enhanced entity and relation extraction.
We regard visual representation as pluggable visual prefix to guide the textual representation for error insensitive forecasting decision.
Experiments on three benchmark datasets demonstrate the effectiveness of our method, and achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-05-07T02:10:55Z) - Parallel Instance Query Network for Named Entity Recognition [73.30174490672647]
Named entity recognition (NER) is a fundamental task in natural language processing.
Recent works treat named entity recognition as a reading comprehension task, constructing type-specific queries manually to extract entities.
We propose Parallel Instance Query Network (PIQN), which sets up global and learnable instance queries to extract entities in a parallel manner.
arXiv Detail & Related papers (2022-03-20T13:01:25Z) - Entity Linking via Dual and Cross-Attention Encoders [16.23946458604865]
We propose a dual-encoder entity retrieval system that learns mention and entity representations in the same space.
We then rerank the entities by using a cross-attention encoder over the target mention and each of the candidate entities.
We achieve state-of-art results on TACKBP-2010 dataset, with 92.05% accuracy.
arXiv Detail & Related papers (2020-04-07T17:28:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.