OAG-BERT: Pre-train Heterogeneous Entity-augmented Academic Language
Model
- URL: http://arxiv.org/abs/2103.02410v1
- Date: Wed, 3 Mar 2021 14:00:57 GMT
- Title: OAG-BERT: Pre-train Heterogeneous Entity-augmented Academic Language
Model
- Authors: Xiao Liu, Da Yin, Xingjian Zhang, Kai Su, Kan Wu, Hongxia Yang, Jie
Tang
- Abstract summary: OAG-BERT integrates massive heterogeneous entities including paper, author, concept, venue, and affiliation.
We develop novel pre-training strategies including heterogeneous entity type embedding, entity-aware 2D positional encoding, and span-aware entity masking.
OAG-BERT has been deployed to multiple real-world applications, such as reviewer recommendations for NSFC (National Nature Science Foundation of China) and paper tagging in the AMiner system.
- Score: 45.419270950610624
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To enrich language models with domain knowledge is crucial but difficult.
Based on the world's largest public academic graph Open Academic Graph (OAG),
we pre-train an academic language model, namely OAG-BERT, which integrates
massive heterogeneous entities including paper, author, concept, venue, and
affiliation. To better endow OAG-BERT with the ability to capture entity
information, we develop novel pre-training strategies including heterogeneous
entity type embedding, entity-aware 2D positional encoding, and span-aware
entity masking. For zero-shot inference, we design a special decoding strategy
to allow OAG-BERT to generate entity names from scratch. We evaluate the
OAG-BERT on various downstream academic tasks, including NLP benchmarks,
zero-shot entity inference, heterogeneous graph link prediction, and author
name disambiguation. Results demonstrate the effectiveness of the proposed
pre-training approach to both comprehending academic texts and modeling
knowledge from heterogeneous entities. OAG-BERT has been deployed to multiple
real-world applications, such as reviewer recommendations for NSFC (National
Nature Science Foundation of China) and paper tagging in the AMiner system. It
is also available to the public through the CogDL package.
Related papers
- JEL: Applying End-to-End Neural Entity Linking in JPMorgan Chase [3.4311229392863467]
We propose a novel end-to-end neural entity linking model (JEL) that uses minimal context information and a margin loss to generate entity embeddings.
We show that JEL achieves the state-of-the-art performance to link mentions of company names in financial news with entities in our knowledge graph.
arXiv Detail & Related papers (2024-11-05T00:46:25Z) - Language Models are Graph Learners [70.14063765424012]
Language Models (LMs) are challenging the dominance of domain-specific models, including Graph Neural Networks (GNNs) and Graph Transformers (GTs)
We propose a novel approach that empowers off-the-shelf LMs to achieve performance comparable to state-of-the-art GNNs on node classification tasks.
arXiv Detail & Related papers (2024-10-03T08:27:54Z) - Entity Disambiguation via Fusion Entity Decoding [68.77265315142296]
We propose an encoder-decoder model to disambiguate entities with more detailed entity descriptions.
We observe +1.5% improvements in end-to-end entity linking in the GERBIL benchmark compared with EntQA.
arXiv Detail & Related papers (2024-04-02T04:27:54Z) - GiGaMAE: Generalizable Graph Masked Autoencoder via Collaborative Latent
Space Reconstruction [76.35904458027694]
Masked autoencoder models lack good generalization ability on graph data.
We propose a novel graph masked autoencoder framework called GiGaMAE.
Our results will shed light on the design of foundation models on graph-structured data.
arXiv Detail & Related papers (2023-08-18T16:30:51Z) - Understand the Dynamic World: An End-to-End Knowledge Informed Framework
for Open Domain Entity State Tracking [15.421012879083463]
Open domain entity state tracking aims to predict reasonable state changes of entities (i.e., [attribute] of [entity] was [before_state] and [after_state] afterwards) given the action descriptions.
It's challenging as the model needs to predict an arbitrary number of entity state changes caused by the action while most of the entities are implicitly relevant to the actions and their attributes as well as states are from open vocabularies.
We propose a novel end-to-end Knowledge Informed framework for open domain Entity State Tracking, namely KIEST, which explicitly retrieves the relevant entities and attributes from
arXiv Detail & Related papers (2023-04-26T22:45:30Z) - Knowledge Graph Augmented Network Towards Multiview Representation
Learning for Aspect-based Sentiment Analysis [96.53859361560505]
We propose a knowledge graph augmented network (KGAN) to incorporate external knowledge with explicitly syntactic and contextual information.
KGAN captures the sentiment feature representations from multiple perspectives, i.e., context-, syntax- and knowledge-based.
Experiments on three popular ABSA benchmarks demonstrate the effectiveness and robustness of our KGAN.
arXiv Detail & Related papers (2022-01-13T08:25:53Z) - Knowledge-Rich Self-Supervised Entity Linking [58.838404666183656]
Knowledge-RIch Self-Supervision ($tt KRISSBERT$) is a universal entity linker for four million UMLS entities.
Our approach subsumes zero-shot and few-shot methods, and can easily incorporate entity descriptions and gold mention labels if available.
Without using any labeled information, our method produces $tt KRISSBERT$, a universal entity linker for four million UMLS entities.
arXiv Detail & Related papers (2021-12-15T05:05:12Z) - GN-Transformer: Fusing Sequence and Graph Representation for Improved
Code Summarization [0.0]
We propose a novel method, GN-Transformer, to learn end-to-end on a fused sequence and graph modality.
The proposed methods achieve state-of-the-art performance in two code summarization datasets and across three automatic code summarization metrics.
arXiv Detail & Related papers (2021-11-17T02:51:37Z) - KI-BERT: Infusing Knowledge Context for Better Language and Domain
Understanding [0.0]
We propose a technique to infuse knowledge context from knowledge graphs for conceptual and ambiguous entities into models based on transformer architecture.
Our novel technique project knowledge graph embedding in the homogeneous vector-space, introduces new token-types for entities, align entity position ids, and a selective attention mechanism.
We take BERT as a baseline model and implement "KnowledgeInfused BERT" by infusing knowledge context from ConceptNet and WordNet.
arXiv Detail & Related papers (2021-04-09T16:15:31Z) - Neural Entity Summarization with Joint Encoding and Weak Supervision [29.26714907483851]
In knowledge graphs, an entity is often described by a large number of triple facts.
Existing solutions to entitymarization are mainly unsupervised.
We present a supervised approach that is based on our novel neural model.
arXiv Detail & Related papers (2020-05-01T00:14:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.