AMR Parsing with Action-Pointer Transformer
- URL: http://arxiv.org/abs/2104.14674v1
- Date: Thu, 29 Apr 2021 22:01:41 GMT
- Title: AMR Parsing with Action-Pointer Transformer
- Authors: Jiawei Zhou, Tahira Naseem, Ram\'on Fernandez Astudillo, Radu Florian
- Abstract summary: We propose a transition-based system that combines hard-attention over sentences with a target-side action pointer mechanism.
We show that our action-pointer approach leads to increased expressiveness and attains large gains against the best transition-based AMR.
- Score: 18.382148821100152
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Abstract Meaning Representation parsing is a sentence-to-graph prediction
task where target nodes are not explicitly aligned to sentence tokens. However,
since graph nodes are semantically based on one or more sentence tokens,
implicit alignments can be derived. Transition-based parsers operate over the
sentence from left to right, capturing this inductive bias via alignments at
the cost of limited expressiveness. In this work, we propose a transition-based
system that combines hard-attention over sentences with a target-side action
pointer mechanism to decouple source tokens from node representations and
address alignments. We model the transitions as well as the pointer mechanism
through straightforward modifications within a single Transformer architecture.
Parser state and graph structure information are efficiently encoded using
attention heads. We show that our action-pointer approach leads to increased
expressiveness and attains large gains (+1.6 points) against the best
transition-based AMR parser in very similar conditions. While using no graph
re-categorization, our single model yields the second best Smatch score on AMR
2.0 (81.8), which is further improved to 83.4 with silver data and ensemble
decoding.
Related papers
- Graph as Point Set [31.448841287258116]
This paper introduces a novel graph-to-set conversion method that transforms interconnected nodes into a set of independent points.
It enables using set encoders to learn from graphs, thereby significantly expanding the design space of Graph Neural Networks.
To demonstrate the effectiveness of our approach, we introduce Point Set Transformer (PST), a transformer architecture that accepts a point set converted from a graph as input.
arXiv Detail & Related papers (2024-05-05T02:29:41Z) - Guiding AMR Parsing with Reverse Graph Linearization [45.37129580211495]
We propose a novel Reverse Graph Linearization (RGL) framework for AMR parsing.
RGL defines both default and reverse linearization orders of an AMR graph, where most structures at the back part of the default order appear at the front part of the reversed order and vice versa.
Our analysis shows that our proposed method significantly mitigates the problem of structure loss accumulation, outperforming the previously best AMR parsing model by 0.8 and 0.5 Smatch scores on the AMR 2.0 and AMR 3.0 dataset, respectively.
arXiv Detail & Related papers (2023-10-13T05:03:13Z) - Discrete Graph Auto-Encoder [52.50288418639075]
We introduce a new framework named Discrete Graph Auto-Encoder (DGAE)
We first use a permutation-equivariant auto-encoder to convert graphs into sets of discrete latent node representations.
In the second step, we sort the sets of discrete latent representations and learn their distribution with a specifically designed auto-regressive model.
arXiv Detail & Related papers (2023-06-13T12:40:39Z) - AGFormer: Efficient Graph Representation with Anchor-Graph Transformer [95.1825252182316]
We propose a novel graph Transformer architecture, termed Anchor Graph Transformer (AGFormer)
AGFormer first obtains some representative anchors and then converts node-to-node message passing into anchor-to-anchor and anchor-to-node message passing process.
Extensive experiments on several benchmark datasets demonstrate the effectiveness and benefits of proposed AGFormer.
arXiv Detail & Related papers (2023-05-12T14:35:42Z) - Paraformer: Fast and Accurate Parallel Transformer for
Non-autoregressive End-to-End Speech Recognition [62.83832841523525]
We propose a fast and accurate parallel transformer, termed Paraformer.
It accurately predicts the number of output tokens and extract hidden variables.
It can attain comparable performance to the state-of-the-art AR transformer, with more than 10x speedup.
arXiv Detail & Related papers (2022-06-16T17:24:14Z) - GN-Transformer: Fusing Sequence and Graph Representation for Improved
Code Summarization [0.0]
We propose a novel method, GN-Transformer, to learn end-to-end on a fused sequence and graph modality.
The proposed methods achieve state-of-the-art performance in two code summarization datasets and across three automatic code summarization metrics.
arXiv Detail & Related papers (2021-11-17T02:51:37Z) - Structure-aware Fine-tuning of Sequence-to-sequence Transformers for
Transition-based AMR Parsing [20.67024416678313]
We explore the integration of general pre-trained sequence-to-sequence language models and a structure-aware transition-based approach.
We propose a simplified transition set, designed to better exploit pre-trained language models for structured fine-tuning.
We show that the proposed parsing architecture retains the desirable properties of previous transition-based approaches, while being simpler and reaching the new state of the art for AMR 2.0, without the need for graph re-categorization.
arXiv Detail & Related papers (2021-10-29T04:36:31Z) - Levi Graph AMR Parser using Heterogeneous Attention [17.74208462902158]
This paper presents a novel approach to AMR parsing by combining heterogeneous data (tokens, concepts, labels) as one input to a transformer to learn attention.
Although our models use significantly fewer parameters than the previous state-of-the-art graph, they show similar or better accuracy on AMR 2.0 and 3.0.
arXiv Detail & Related papers (2021-07-09T00:06:17Z) - Span Pointer Networks for Non-Autoregressive Task-Oriented Semantic
Parsing [55.97957664897004]
An effective recipe for building seq2seq, non-autoregressive, task-orienteds to map utterances to semantic frames proceeds in three steps.
These models are typically bottlenecked by length prediction.
In our work, we propose non-autoregressives which shift the decoding task from text generation to span prediction.
arXiv Detail & Related papers (2021-04-15T07:02:35Z) - Fast End-to-End Speech Recognition via a Non-Autoregressive Model and
Cross-Modal Knowledge Transferring from BERT [72.93855288283059]
We propose a non-autoregressive speech recognition model called LASO (Listen Attentively, and Spell Once)
The model consists of an encoder, a decoder, and a position dependent summarizer (PDS)
arXiv Detail & Related papers (2021-02-15T15:18:59Z) - A Differentiable Relaxation of Graph Segmentation and Alignment for AMR
Parsing [75.36126971685034]
We treat alignment and segmentation as latent variables in our model and induce them as part of end-to-end training.
Our method also approaches that of a model that relies on citetLyu2018AMRPA's segmentation rules, which were hand-crafted to handle individual AMR constructions.
arXiv Detail & Related papers (2020-10-23T21:22:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.