Explanation Graph Generation via Generative Pre-training over Synthetic
Graphs
- URL: http://arxiv.org/abs/2306.00652v1
- Date: Thu, 1 Jun 2023 13:20:22 GMT
- Title: Explanation Graph Generation via Generative Pre-training over Synthetic
Graphs
- Authors: Han Cui, Shangzhan Li, Yu Zhang and Qi Shi
- Abstract summary: The generation of explanation graphs is a significant task that aims to produce explanation graphs in response to user input.
Current research commonly fine-tunes a text-based pre-trained language model on a small downstream dataset that is annotated with labeled graphs.
We propose a novel pre-trained framework EG3P(for Explanation Graph Generation via Generative Pre-training over synthetic graphs) for the explanation graph generation task.
- Score: 6.25568933262682
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The generation of explanation graphs is a significant task that aims to
produce explanation graphs in response to user input, revealing the internal
reasoning process. This task is challenging due to the significant discrepancy
between unstructured user queries and structured explanation graphs. Current
research commonly fine-tunes a text-based pre-trained language model on a small
downstream dataset that is annotated with labeled graphs. However, due to the
limited scale of available datasets, this approach may prove to be insufficient
in bridging the gap between natural language text and structured graphs. In
this paper, to alleviate the above limitations, we propose a novel pre-trained
framework EG3P(for Explanation Graph Generation via Generative Pre-training
over synthetic graphs) for the explanation graph generation task. Specifically,
we first propose a text-to-graph generative task to pre-train the model with
the goal of bridging the text-graph gap. Additionally, we propose an automatic
corpus synthesis strategy for synthesizing a large scale of high-quality
corpus, reducing the reliance on costly manual annotation methods. Experimental
results on ExplaGraphs show the effectiveness of EG3P that our model surpasses
all baseline systems with remarkable margins. Besides, further analysis
demonstrates that EG3P is able to generate better explanation graphs on actual
reasoning tasks such as CommonsenseQA and OpenbookQA.
Related papers
- G-Retriever: Retrieval-Augmented Generation for Textual Graph Understanding and Question Answering [61.93058781222079]
We develop a flexible question-answering framework targeting real-world textual graphs.
We introduce the first retrieval-augmented generation (RAG) approach for general textual graphs.
G-Retriever performs RAG over a graph by formulating this task as a Prize-Collecting Steiner Tree optimization problem.
arXiv Detail & Related papers (2024-02-12T13:13:04Z) - When Graph Data Meets Multimodal: A New Paradigm for Graph Understanding
and Reasoning [54.84870836443311]
The paper presents a new paradigm for understanding and reasoning about graph data by integrating image encoding and multimodal technologies.
This approach enables the comprehension of graph data through an instruction-response format, utilizing GPT-4V's advanced capabilities.
The study evaluates this paradigm on various graph types, highlighting the model's strengths and weaknesses, particularly in Chinese OCR performance and complex reasoning tasks.
arXiv Detail & Related papers (2023-12-16T08:14:11Z) - Form follows Function: Text-to-Text Conditional Graph Generation based
on Functional Requirements [36.00630198983932]
This work focuses on the novel problem setting of generating graphs conditioned on a description of the graph's functional requirements in a downstream task.
We pose the problem as a text-to-text generation problem and focus on the approach of fine-tuning a pretrained large language model (LLM) to generate graphs.
arXiv Detail & Related papers (2023-11-01T11:12:02Z) - Let There Be Order: Rethinking Ordering in Autoregressive Graph
Generation [6.422073551199993]
Conditional graph generation tasks involve training a model to generate a graph given a set of input conditions.
Many previous studies employ autoregressive models to incrementally generate graph components such as nodes and edges.
As graphs typically lack a natural ordering among their components, converting a graph into a sequence of tokens is not straightforward.
arXiv Detail & Related papers (2023-05-24T20:52:34Z) - Explanation Graph Generation via Pre-trained Language Models: An
Empirical Study with Contrastive Learning [84.35102534158621]
We study pre-trained language models that generate explanation graphs in an end-to-end manner.
We propose simple yet effective ways of graph perturbations via node and edge edit operations.
Our methods lead to significant improvements in both structural and semantic accuracy of explanation graphs.
arXiv Detail & Related papers (2022-04-11T00:58:27Z) - Promoting Graph Awareness in Linearized Graph-to-Text Generation [72.83863719868364]
We study the ability of linearized models to encode local graph structures.
Our findings motivate solutions to enrich the quality of models' implicit graph encodings.
We find that these denoising scaffolds lead to substantial improvements in downstream generation in low-resource settings.
arXiv Detail & Related papers (2020-12-31T18:17:57Z) - Neural Language Modeling for Contextualized Temporal Graph Generation [49.21890450444187]
This paper presents the first study on using large-scale pre-trained language models for automated generation of an event-level temporal graph for a document.
arXiv Detail & Related papers (2020-10-20T07:08:00Z) - Graph Information Bottleneck for Subgraph Recognition [103.37499715761784]
We propose a framework of Graph Information Bottleneck (GIB) for the subgraph recognition problem in deep graph learning.
Under this framework, one can recognize the maximally informative yet compressive subgraph, named IB-subgraph.
We evaluate the properties of the IB-subgraph in three application scenarios: improvement of graph classification, graph interpretation and graph denoising.
arXiv Detail & Related papers (2020-10-12T09:32:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.