Graph-to-Sequence Neural Machine Translation
- URL: http://arxiv.org/abs/2009.07489v1
- Date: Wed, 16 Sep 2020 06:28:58 GMT
- Title: Graph-to-Sequence Neural Machine Translation
- Authors: Sufeng Duan, Hai Zhao and Rui Wang
- Abstract summary: We propose a graph-based SAN-based NMT model called Graph-Transformer.
Subgraphs are put into different groups according to their orders, and every group of subgraphs respectively reflect different levels of dependency between words.
Our method can effectively boost the Transformer with an improvement of 1.1 BLEU points on WMT14 English-German dataset and 1.0 BLEU points on IWSLT14 German-English dataset.
- Score: 79.0617920270817
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural machine translation (NMT) usually works in a seq2seq learning way by
viewing either source or target sentence as a linear sequence of words, which
can be regarded as a special case of graph, taking words in the sequence as
nodes and relationships between words as edges. In the light of the current NMT
models more or less capture graph information among the sequence in a latent
way, we present a graph-to-sequence model facilitating explicit graph
information capturing. In detail, we propose a graph-based SAN-based NMT model
called Graph-Transformer by capturing information of subgraphs of different
orders in every layers. Subgraphs are put into different groups according to
their orders, and every group of subgraphs respectively reflect different
levels of dependency between words. For fusing subgraph representations, we
empirically explore three methods which weight different groups of subgraphs of
different orders. Results of experiments on WMT14 English-German and IWSLT14
German-English show that our method can effectively boost the Transformer with
an improvement of 1.1 BLEU points on WMT14 English-German dataset and 1.0 BLEU
points on IWSLT14 German-English dataset.
Related papers
- On Pre-training of Multimodal Language Models Customized for Chart Understanding [83.99377088129282]
This paper explores the training processes necessary to improve MLLMs' comprehension of charts.
We introduce CHOPINLLM, an MLLM tailored for in-depth chart comprehension.
arXiv Detail & Related papers (2024-07-19T17:58:36Z) - Parameter-Efficient Tuning Large Language Models for Graph Representation Learning [62.26278815157628]
We introduce Graph-aware.
Efficient Fine-Tuning - GPEFT, a novel approach for efficient graph representation learning.
We use a graph neural network (GNN) to encode structural information from neighboring nodes into a graph prompt.
We validate our approach through comprehensive experiments conducted on 8 different text-rich graphs, observing an average improvement of 2% in hit@1 and Mean Reciprocal Rank (MRR) in link prediction evaluations.
arXiv Detail & Related papers (2024-04-28T18:36:59Z) - Graph Mixup with Soft Alignments [49.61520432554505]
We study graph data augmentation by mixup, which has been used successfully on images.
We propose S-Mixup, a simple yet effective mixup method for graph classification by soft alignments.
arXiv Detail & Related papers (2023-06-11T22:04:28Z) - FACTUAL: A Benchmark for Faithful and Consistent Textual Scene Graph
Parsing [66.70054075041487]
Existing scene graphs that convert image captions into scene graphs often suffer from two types of errors.
First, the generated scene graphs fail to capture the true semantics of the captions or the corresponding images, resulting in a lack of faithfulness.
Second, the generated scene graphs have high inconsistency, with the same semantics represented by different annotations.
arXiv Detail & Related papers (2023-05-27T15:38:31Z) - Dependency Graph-to-String Statistical Machine Translation [40.91501320179243]
Source graphs are constructed from dependency trees with extra links so that non-syntactic phrases are connected.
Inspired by phrase-based models, we first introduce a translation model which segments a graph into a sequence of disjoint subgraphs and generates a translation by combining subgraph translations left-to-right using beam search.
arXiv Detail & Related papers (2021-03-20T04:20:56Z) - To Understand Representation of Layer-aware Sequence Encoders as
Multi-order-graph [45.51774590045651]
We propose a unified explanation of representation for layer-aware neural sequence encoders.
Our proposed MoG explanation allows to precisely observe every step of the generation of representation.
We also propose a graph-based self-attention network empowered Graph-Transformer.
arXiv Detail & Related papers (2021-01-16T08:12:03Z) - Motif-Driven Contrastive Learning of Graph Representations [32.03481571304036]
We propose to learn graph motifs, which are frequently-occurring subgraph patterns, for better subgraph sampling.
By pre-training on the ogbg-molhiv dataset with MICRO-Graph, the pre-trained GNN achieves 2.04% ROC-AUC average performance enhancement.
arXiv Detail & Related papers (2020-12-23T08:10:19Z) - Document Graph for Neural Machine Translation [42.13593962963306]
We show that a document can be represented as a graph that connects relevant contexts regardless of their distances.
Experiments on various NMT benchmarks, including IWSLT English-French, Chinese-English, WMT English-German and Opensubtitle English-Russian, demonstrate that using document graphs can significantly improve the translation quality.
arXiv Detail & Related papers (2020-12-07T06:48:59Z) - Sequential Graph Convolutional Network for Active Learning [53.99104862192055]
We propose a novel pool-based Active Learning framework constructed on a sequential Graph Convolution Network (GCN)
With a small number of randomly sampled images as seed labelled examples, we learn the parameters of the graph to distinguish labelled vs unlabelled nodes.
We exploit these characteristics of GCN to select the unlabelled examples which are sufficiently different from labelled ones.
arXiv Detail & Related papers (2020-06-18T00:55:10Z) - Tensor Graph Convolutional Networks for Text Classification [17.21683037822181]
Graph-based neural networks exhibit some excellent properties, such as ability capturing global information.
In this paper, we investigate graph-based neural networks for text classification problem.
arXiv Detail & Related papers (2020-01-12T14:28:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.