ALBERT with Knowledge Graph Encoder Utilizing Semantic Similarity for
Commonsense Question Answering
- URL: http://arxiv.org/abs/2211.07065v1
- Date: Mon, 14 Nov 2022 01:39:26 GMT
- Title: ALBERT with Knowledge Graph Encoder Utilizing Semantic Similarity for
Commonsense Question Answering
- Authors: Byeongmin Choi, YongHyun Lee, Yeunwoong Kyung and Eunchan Kim
- Abstract summary: We propose to use the latest pre-trained language model a lite bidirectional encoder representations from transformers (ALBERT) with knowledge graph information extraction technique.
We show that our proposed model can achieve better performance than existing KagNet and MHGRN models in CommonsenseQA dataset.
- Score: 1.7205106391379026
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recently, pre-trained language representation models such as bidirectional
encoder representations from transformers (BERT) have been performing well in
commonsense question answering (CSQA). However, there is a problem that the
models do not directly use explicit information of knowledge sources existing
outside. To augment this, additional methods such as knowledge-aware graph
network (KagNet) and multi-hop graph relation network (MHGRN) have been
proposed. In this study, we propose to use the latest pre-trained language
model a lite bidirectional encoder representations from transformers (ALBERT)
with knowledge graph information extraction technique. We also propose to
applying the novel method, schema graph expansion to recent language models.
Then, we analyze the effect of applying knowledge graph-based knowledge
extraction techniques to recent pre-trained language models and confirm that
schema graph expansion is effective in some extent. Furthermore, we show that
our proposed model can achieve better performance than existing KagNet and
MHGRN models in CommonsenseQA dataset.
Related papers
- ReasoningLM: Enabling Structural Subgraph Reasoning in Pre-trained
Language Models for Question Answering over Knowledge Graph [142.42275983201978]
We propose a subgraph-aware self-attention mechanism to imitate the GNN for performing structured reasoning.
We also adopt an adaptation tuning strategy to adapt the model parameters with 20,000 subgraphs with synthesized questions.
Experiments show that ReasoningLM surpasses state-of-the-art models by a large margin, even with fewer updated parameters and less training data.
arXiv Detail & Related papers (2023-12-30T07:18:54Z) - Hi-ArG: Exploring the Integration of Hierarchical Argumentation Graphs
in Language Pretraining [62.069374456021016]
We propose the Hierarchical Argumentation Graph (Hi-ArG), a new structure to organize arguments.
We also introduce two approaches to exploit Hi-ArG, including a text-graph multi-modal model GreaseArG and a new pre-training framework augmented with graph information.
arXiv Detail & Related papers (2023-12-01T19:03:38Z) - LUKE-Graph: A Transformer-based Approach with Gated Relational Graph
Attention for Cloze-style Reading Comprehension [13.173307471333619]
We propose the LUKE-Graph, a model that builds a heterogeneous graph based on the intuitive relationships between entities in a document.
We then use the Attention reading (RGAT) to fuse the graph's reasoning information and the contextual representation encoded by the pre-trained LUKE model.
Experimental results demonstrate that the LUKE-Graph achieves state-of-the-art performance with commonsense reasoning.
arXiv Detail & Related papers (2023-03-12T14:31:44Z) - Text Representation Enrichment Utilizing Graph based Approaches: Stock
Market Technical Analysis Case Study [0.0]
We propose a transductive hybrid approach composed of an unsupervised node representation learning model followed by a node classification/edge prediction model.
The proposed model is developed to classify stock market technical analysis reports, which to our knowledge is the first work in this domain.
arXiv Detail & Related papers (2022-11-29T11:26:08Z) - KGLM: Integrating Knowledge Graph Structure in Language Models for Link
Prediction [0.0]
We introduce a new entity/relation embedding layer that learns to differentiate distinctive entity and relation types.
We show that further pre-training the language models with this additional embedding layer using the triples extracted from the knowledge graph, followed by the standard fine-tuning phase sets a new state-of-the-art performance for the link prediction task on the benchmark datasets.
arXiv Detail & Related papers (2022-11-04T20:38:12Z) - An Empirical Study of Retrieval-enhanced Graph Neural Networks [48.99347386689936]
Graph Neural Networks (GNNs) are effective tools for graph representation learning.
We propose a retrieval-enhanced scheme called GRAPHRETRIEVAL, which is agnostic to the choice of graph neural network models.
We conduct comprehensive experiments over 13 datasets, and we observe that GRAPHRETRIEVAL is able to reach substantial improvements over existing GNNs.
arXiv Detail & Related papers (2022-06-01T09:59:09Z) - GAP: A Graph-aware Language Model Framework for Knowledge Graph-to-Text
Generation [3.593955557310285]
Recent improvements in KG-to-text generation are due to auxiliary pre-training tasks designed to give the fine-tune task a boost in performance.
Here, we demonstrate that by fusing graph-aware elements into existing pre-trained language models, we are able to outperform state-of-the-art models and close the gap imposed by additional pre-training tasks.
arXiv Detail & Related papers (2022-04-13T23:53:37Z) - KGRefiner: Knowledge Graph Refinement for Improving Accuracy of
Translational Link Prediction Methods [4.726777092009553]
This paper proposes a method for refining the knowledge graph.
It makes the knowledge graph more informative, and link prediction operations can be performed more accurately.
Our experiments show that our method can significantly increase the performance of translational link prediction methods.
arXiv Detail & Related papers (2021-06-27T13:32:39Z) - GraphFormers: GNN-nested Transformers for Representation Learning on
Textual Graph [53.70520466556453]
We propose GraphFormers, where layerwise GNN components are nested alongside the transformer blocks of language models.
With the proposed architecture, the text encoding and the graph aggregation are fused into an iterative workflow.
In addition, a progressive learning strategy is introduced, where the model is successively trained on manipulated data and original data to reinforce its capability of integrating information on graph.
arXiv Detail & Related papers (2021-05-06T12:20:41Z) - Learning Intents behind Interactions with Knowledge Graph for
Recommendation [93.08709357435991]
Knowledge graph (KG) plays an increasingly important role in recommender systems.
Existing GNN-based models fail to identify user-item relation at a fine-grained level of intents.
We propose a new model, Knowledge Graph-based Intent Network (KGIN)
arXiv Detail & Related papers (2021-02-14T03:21:36Z) - Distilling Knowledge from Graph Convolutional Networks [146.71503336770886]
Existing knowledge distillation methods focus on convolutional neural networks (CNNs)
We propose the first dedicated approach to distilling knowledge from a pre-trained graph convolutional network (GCN) model.
We show that our method achieves the state-of-the-art knowledge distillation performance for GCN models.
arXiv Detail & Related papers (2020-03-23T18:23:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.