Pairwise Learning for Name Disambiguation in Large-Scale Heterogeneous
Academic Networks
- URL: http://arxiv.org/abs/2008.13099v4
- Date: Wed, 20 Jan 2021 12:31:45 GMT
- Title: Pairwise Learning for Name Disambiguation in Large-Scale Heterogeneous
Academic Networks
- Authors: Qingyun Sun, Hao Peng, Jianxin Li, Senzhang Wang, Xiangyu Dong,
Liangxuan Zhao, Philip S. Yu and Lifang He
- Abstract summary: We introduce Multi-view Attention-based Pairwise Recurrent Neural Network (MA-PairRNN) to solve the name disambiguation problem.
MA-PairRNN combines heterogeneous graph embedding learning and pairwise similarity learning into a framework.
Results on two real-world datasets demonstrate that our framework has a significant and consistent improvement of performance on the name disambiguation task.
- Score: 81.00481125272098
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Name disambiguation aims to identify unique authors with the same name.
Existing name disambiguation methods always exploit author attributes to
enhance disambiguation results. However, some discriminative author attributes
(e.g., email and affiliation) may change because of graduation or job-hopping,
which will result in the separation of the same author's papers in digital
libraries. Although these attributes may change, an author's co-authors and
research topics do not change frequently with time, which means that papers
within a period have similar text and relation information in the academic
network. Inspired by this idea, we introduce Multi-view Attention-based
Pairwise Recurrent Neural Network (MA-PairRNN) to solve the name disambiguation
problem. We divided papers into small blocks based on discriminative author
attributes and blocks of the same author will be merged according to pairwise
classification results of MA-PairRNN. MA-PairRNN combines heterogeneous graph
embedding learning and pairwise similarity learning into a framework. In
addition to attribute and structure information, MA-PairRNN also exploits
semantic information by meta-path and generates node representation in an
inductive way, which is scalable to large graphs. Furthermore, a semantic-level
attention mechanism is adopted to fuse multiple meta-path based
representations. A Pseudo-Siamese network consisting of two RNNs takes two
paper sequences in publication time order as input and outputs their
similarity. Results on two real-world datasets demonstrate that our framework
has a significant and consistent improvement of performance on the name
disambiguation task. It was also demonstrated that MA-PairRNN can perform well
with a small amount of training data and have better generalization ability
across different research areas.
Related papers
- A Generative Approach for Wikipedia-Scale Visual Entity Recognition [56.55633052479446]
We address the task of mapping a given query image to one of the 6 million existing entities in Wikipedia.
We introduce a novel Generative Entity Recognition framework, which learns to auto-regressively decode a semantic and discriminative code'' identifying the target entity.
arXiv Detail & Related papers (2024-03-04T13:47:30Z) - Language Models As Semantic Indexers [78.83425357657026]
We introduce LMIndexer, a self-supervised framework to learn semantic IDs with a generative language model.
We show the high quality of the learned IDs and demonstrate their effectiveness on three tasks including recommendation, product search, and document retrieval.
arXiv Detail & Related papers (2023-10-11T18:56:15Z) - Named Entity Recognition via Machine Reading Comprehension: A Multi-Task
Learning Approach [50.12455129619845]
Named Entity Recognition (NER) aims to extract and classify entity mentions in the text into pre-defined types.
We propose to incorporate the label dependencies among entity types into a multi-task learning framework for better MRC-based NER.
arXiv Detail & Related papers (2023-09-20T03:15:05Z) - Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - Author Name Disambiguation via Heterogeneous Network Embedding from
Structural and Semantic Perspectives [13.266320447769564]
Name ambiguity is common in academic digital libraries, such as multiple authors having the same name.
The proposed method is mainly based on representation learning for heterogeneous networks and clustering.
The semantic representation is generated using NLP tools.
arXiv Detail & Related papers (2022-12-24T11:22:34Z) - The Fellowship of the Authors: Disambiguating Names from Social Network
Context [2.3605348648054454]
Authority lists with extensive textual descriptions for each entity are lacking and ambiguous named entities.
We combine BERT-based mention representations with a variety of graph induction strategies and experiment with supervised and unsupervised cluster inference methods.
We find that in-domain language model pretraining can significantly improve mention representations, especially for larger corpora.
arXiv Detail & Related papers (2022-08-31T21:51:55Z) - NameRec*: Highly Accurate and Fine-grained Person Name Recognition [11.43547342030705]
NameRec* task aims to do highly accurate and fine-grained person name recognition.
CogNN fully explores the intra-sentence context and rich training signals of name forms.
IsBERT has an overlapped input processor, and an inter-sentence encoder with bidirectional overlapped contextual embedding learning.
arXiv Detail & Related papers (2021-03-21T10:35:04Z) - Effect of forename string on author name disambiguation [8.160343645537106]
Author forenames are used to decide which name instances are disambiguated together and how much they are likely to refer to the same author.
This study assesses the contributions of forenames in author name disambiguation using multiple labeled datasets.
arXiv Detail & Related papers (2021-02-05T15:54:11Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - Joint Semantic Analysis with Document-Level Cross-Task Coherence Rewards [13.753240692520098]
We present a neural network architecture for joint coreference resolution and semantic role labeling for English.
We use reinforcement learning to encourage global coherence over the document and between semantic annotations.
This leads to improvements on both tasks in multiple datasets from different domains.
arXiv Detail & Related papers (2020-10-12T09:36:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.