AMR Parsing via Graph-Sequence Iterative Inference
- URL: http://arxiv.org/abs/2004.05572v2
- Date: Wed, 29 Apr 2020 04:01:44 GMT
- Title: AMR Parsing via Graph-Sequence Iterative Inference
- Authors: Deng Cai and Wai Lam
- Abstract summary: We propose a new end-to-end model that treats AMR parsing as a series of dual decisions on the input sequence and the incrementally constructed graph.
We show that the answers to these two questions are mutually causalities.
We design a model based on iterative inference that helps achieve better answers in both perspectives, leading to greatly improved parsing accuracy.
- Score: 62.85003739964878
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new end-to-end model that treats AMR parsing as a series of dual
decisions on the input sequence and the incrementally constructed graph. At
each time step, our model performs multiple rounds of attention, reasoning, and
composition that aim to answer two critical questions: (1) which part of the
input \textit{sequence} to abstract; and (2) where in the output \textit{graph}
to construct the new concept. We show that the answers to these two questions
are mutually causalities. We design a model based on iterative inference that
helps achieve better answers in both perspectives, leading to greatly improved
parsing accuracy. Our experimental results significantly outperform all
previously reported \textsc{Smatch} scores by large margins. Remarkably,
without the help of any large-scale pre-trained language model (e.g., BERT),
our model already surpasses previous state-of-the-art using BERT. With the help
of BERT, we can push the state-of-the-art results to 80.2\% on LDC2017T10 (AMR
2.0) and 75.4\% on LDC2014T12 (AMR 1.0).
Related papers
- Less is More: One-shot Subgraph Reasoning on Large-scale Knowledge Graphs [49.547988001231424]
We propose the one-shot-subgraph link prediction to achieve efficient and adaptive prediction.
Design principle is that, instead of directly acting on the whole KG, the prediction procedure is decoupled into two steps.
We achieve promoted efficiency and leading performances on five large-scale benchmarks.
arXiv Detail & Related papers (2024-03-15T12:00:12Z) - MinPrompt: Graph-based Minimal Prompt Data Augmentation for Few-shot Question Answering [64.6741991162092]
We present MinPrompt, a minimal data augmentation framework for open-domain question answering.
We transform the raw text into a graph structure to build connections between different factual sentences.
We then apply graph algorithms to identify the minimal set of sentences needed to cover the most information in the raw text.
We generate QA pairs based on the identified sentence subset and train the model on the selected sentences to obtain the final model.
arXiv Detail & Related papers (2023-10-08T04:44:36Z) - RGAT: A Deeper Look into Syntactic Dependency Information for
Coreference Resolution [8.017036537163008]
We propose an end-to-end resolution that combines pre-trained BERT with a Syntactic Relation Graph Attention Network (RGAT)
In particular, the RGAT model is first proposed, then used to understand the syntactic dependency graph and learn better task-specific syntactic embeddings.
An integrated architecture incorporating BERT embeddings and syntactic embeddings is constructed to generate blending representations for the downstream task.
arXiv Detail & Related papers (2023-09-10T09:46:38Z) - A Question-Answering Approach to Key Value Pair Extraction from
Form-like Document Images [8.73248722579337]
We present a new question-answering (QA) based key-value pair extraction approach, called KVPFormer.
We propose a coarse-to-fine answer prediction approach to achieve higher answer prediction accuracy.
Our proposed Ours achieves state-of-the-art results on FUNSD and XFUND datasets, outperforming the previous best-performing method by 7.2% and 13.2% in F1 score, respectively.
arXiv Detail & Related papers (2023-04-17T02:55:31Z) - Tokenization Consistency Matters for Generative Models on Extractive NLP
Tasks [54.306234256074255]
We identify the issue of tokenization inconsistency that is commonly neglected in training generative models.
This issue damages the extractive nature of these tasks after the input and output are tokenized inconsistently.
We show that, with consistent tokenization, the model performs better in both in-domain and out-of-domain datasets.
arXiv Detail & Related papers (2022-12-19T23:33:21Z) - Enriching Relation Extraction with OpenIE [70.52564277675056]
Relation extraction (RE) is a sub-discipline of information extraction (IE)
In this work, we explore how recent approaches for open information extraction (OpenIE) may help to improve the task of RE.
Our experiments over two annotated corpora, KnowledgeNet and FewRel, demonstrate the improved accuracy of our enriched models.
arXiv Detail & Related papers (2022-12-19T11:26:23Z) - Ensembling Graph Predictions for AMR Parsing [28.625065956013778]
In many machine learning tasks, models are trained to predict structure data such as graphs.
In this work, we formalize this problem as mining the largest graph that is the most supported by a collection of graph predictions.
We show that the proposed approach can combine the strength of state-of-the-art AMRs to create new predictions that are more accurate than any individual models in five standard benchmark datasets.
arXiv Detail & Related papers (2021-10-18T09:35:39Z) - DYLE: Dynamic Latent Extraction for Abstractive Long-Input Summarization [19.693324720440017]
Transformer-based models have achieved state-of-the-art performance on short text summarization.
We present a new approach for long-input summarization: Dynamic Latent Extraction for Abstractive Summarization.
Our model significantly outperforms the current state-of-the-art, including a 6.21 ROUGE-2 improvement on GovReport and a 2.13 ROUGE-1 improvement on QMSum.
arXiv Detail & Related papers (2021-10-15T15:53:32Z) - Improving AMR Parsing with Sequence-to-Sequence Pre-training [39.33133978535497]
In this paper, we focus on sequence-to-sequence (seq2seq) AMR parsing.
We propose a seq2seq pre-training approach to build pre-trained models in both single and joint way.
Experiments show that both the single and joint pre-trained models significantly improve the performance.
arXiv Detail & Related papers (2020-10-05T04:32:47Z) - Structured Multimodal Attentions for TextVQA [57.71060302874151]
We propose an end-to-end structured multimodal attention (SMA) neural network to mainly solve the first two issues above.
SMA first uses a structural graph representation to encode the object-object, object-text and text-text relationships appearing in the image, and then designs a multimodal graph attention network to reason over it.
Our proposed model outperforms the SoTA models on TextVQA dataset and two tasks of ST-VQA dataset among all models except pre-training based TAP.
arXiv Detail & Related papers (2020-06-01T07:07:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.