R2D2: Relational Text Decoding with Transformers
- URL: http://arxiv.org/abs/2105.04645v1
- Date: Mon, 10 May 2021 19:59:11 GMT
- Title: R2D2: Relational Text Decoding with Transformers
- Authors: Aryan Arbabi, Mingqiu Wang, Laurent El Shafey, Nan Du, Izhak Shafran
- Abstract summary: We propose a novel framework for modeling the interaction between graphical structures and the natural language text associated with their nodes and edges.
Our proposed method utilizes both the graphical structure as well as the sequential nature of the texts.
While the proposed model has wide applications, we demonstrate its capabilities on data-to-text generation tasks.
- Score: 18.137828323277347
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel framework for modeling the interaction between graphical
structures and the natural language text associated with their nodes and edges.
Existing approaches typically fall into two categories. On group ignores the
relational structure by converting them into linear sequences and then utilize
the highly successful Seq2Seq models. The other side ignores the sequential
nature of the text by representing them as fixed-dimensional vectors and apply
graph neural networks. Both simplifications lead to information loss.
Our proposed method utilizes both the graphical structure as well as the
sequential nature of the texts. The input to our model is a set of text
segments associated with the nodes and edges of the graph, which are then
processed with a transformer encoder-decoder model, equipped with a
self-attention mechanism that is aware of the graphical relations between the
nodes containing the segments. This also allows us to use BERT-like models that
are already trained on large amounts of text.
While the proposed model has wide applications, we demonstrate its
capabilities on data-to-text generation tasks. Our approach compares favorably
against state-of-the-art methods in four tasks without tailoring the model
architecture. We also provide an early demonstration in a novel practical
application -- generating clinical notes from the medical entities mentioned
during clinical visits.
Related papers
- Dynamic Relation Transformer for Contextual Text Block Detection [9.644204545582742]
Contextual Text Block Detection is the task of identifying coherent text blocks within the complexity of natural scenes.
Previous methodologies have treated CTBD as either a visual relation extraction challenge within computer vision or as a sequence modeling problem.
We introduce a new framework that frames CTBD as a graph generation problem.
arXiv Detail & Related papers (2024-01-17T14:17:59Z) - DocGraphLM: Documental Graph Language Model for Information Extraction [15.649726614383388]
We introduce DocGraphLM, a framework that combines pre-trained language models with graph semantics.
To achieve this, we propose 1) a joint encoder architecture to represent documents, and 2) a novel link prediction approach to reconstruct document graphs.
Our experiments on three SotA datasets show consistent improvement on IE and QA tasks with the adoption of graph features.
arXiv Detail & Related papers (2024-01-05T14:15:36Z) - An Autoregressive Text-to-Graph Framework for Joint Entity and Relation
Extraction [4.194768796374315]
We propose a novel method for joint entity and relation extraction from unstructured text by framing it as a conditional sequence generation problem.
It generates a linearized graph where nodes represent text spans and edges represent relation triplets.
Our method employs a transformer encoder-decoder architecture with pointing mechanism on a dynamic vocabulary of spans and relation types.
arXiv Detail & Related papers (2024-01-02T18:32:14Z) - GrannGAN: Graph annotation generative adversarial networks [72.66289932625742]
We consider the problem of modelling high-dimensional distributions and generating new examples of data with complex relational feature structure coherent with a graph skeleton.
The model we propose tackles the problem of generating the data features constrained by the specific graph structure of each data point by splitting the task into two phases.
In the first it models the distribution of features associated with the nodes of the given graph, in the second it complements the edge features conditionally on the node features.
arXiv Detail & Related papers (2022-12-01T11:49:07Z) - Improving Generation and Evaluation of Visual Stories via Semantic
Consistency [72.00815192668193]
Given a series of natural language captions, an agent must generate a sequence of images that correspond to the captions.
Prior work has introduced recurrent generative models which outperform synthesis text-to-image models on this task.
We present a number of improvements to prior modeling approaches, including the addition of a dual learning framework.
arXiv Detail & Related papers (2021-05-20T20:42:42Z) - GraphFormers: GNN-nested Transformers for Representation Learning on
Textual Graph [53.70520466556453]
We propose GraphFormers, where layerwise GNN components are nested alongside the transformer blocks of language models.
With the proposed architecture, the text encoding and the graph aggregation are fused into an iterative workflow.
In addition, a progressive learning strategy is introduced, where the model is successively trained on manipulated data and original data to reinforce its capability of integrating information on graph.
arXiv Detail & Related papers (2021-05-06T12:20:41Z) - Promoting Graph Awareness in Linearized Graph-to-Text Generation [72.83863719868364]
We study the ability of linearized models to encode local graph structures.
Our findings motivate solutions to enrich the quality of models' implicit graph encodings.
We find that these denoising scaffolds lead to substantial improvements in downstream generation in low-resource settings.
arXiv Detail & Related papers (2020-12-31T18:17:57Z) - CoSE: Compositional Stroke Embeddings [52.529172734044664]
We present a generative model for complex free-form structures such as stroke-based drawing tasks.
Our approach is suitable for interactive use cases such as auto-completing diagrams.
arXiv Detail & Related papers (2020-06-17T15:22:54Z) - Structure-Augmented Text Representation Learning for Efficient Knowledge
Graph Completion [53.31911669146451]
Human-curated knowledge graphs provide critical supportive information to various natural language processing tasks.
These graphs are usually incomplete, urging auto-completion of them.
graph embedding approaches, e.g., TransE, learn structured knowledge via representing graph elements into dense embeddings.
textual encoding approaches, e.g., KG-BERT, resort to graph triple's text and triple-level contextualized representations.
arXiv Detail & Related papers (2020-04-30T13:50:34Z) - Two-Level Transformer and Auxiliary Coherence Modeling for Improved Text
Segmentation [9.416757363901295]
We introduce a novel supervised model for text segmentation with simple but explicit coherence modeling.
Our model -- a neural architecture consisting of two hierarchically connected Transformer networks -- is a multi-task learning model that couples the sentence-level segmentation objective with the coherence objective that differentiates correct sequences of sentences from corrupt ones.
arXiv Detail & Related papers (2020-01-03T17:06:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.