Neural Data-to-Text Generation with Dynamic Content Planning
- URL: http://arxiv.org/abs/2004.07426v2
- Date: Mon, 20 Apr 2020 03:28:46 GMT
- Title: Neural Data-to-Text Generation with Dynamic Content Planning
- Authors: Kai Chen, Fayuan Li, Baotian Hu, Weihua Peng, Qingcai Chen and Hong Yu
- Abstract summary: We propose a Neural data-to-text generation model with Dynamic content Planning, named NDP for abbreviation.
The NDP can utilize the previously generated text to dynamically select the appropriate entry from the given structured data.
Empirical results show that the NDP superior performance over the state-of-the-art on ROTOWIRE dataset.
- Score: 26.98332458548882
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural data-to-text generation models have achieved significant advancement
in recent years. However, these models have two shortcomings: the generated
texts tend to miss some vital information, and they often generate descriptions
that are not consistent with the structured input data. To alleviate these
problems, we propose a Neural data-to-text generation model with Dynamic
content Planning, named NDP for abbreviation. The NDP can utilize the
previously generated text to dynamically select the appropriate entry from the
given structured data. We further design a reconstruction mechanism with a
novel objective function that can reconstruct the whole entry of the used data
sequentially from the hidden states of the decoder, which aids the accuracy of
the generated text. Empirical results show that the NDP achieves superior
performance over the state-of-the-art on ROTOWIRE dataset, in terms of relation
generation (RG), content selection (CS), content ordering (CO) and BLEU
metrics. The human evaluation result shows that the texts generated by the
proposed NDP are better than the corresponding ones generated by NCP in most of
time. And using the proposed reconstruction mechanism, the fidelity of the
generated text can be further improved significantly.
Related papers
- RegaVAE: A Retrieval-Augmented Gaussian Mixture Variational Auto-Encoder
for Language Modeling [79.56442336234221]
We introduce RegaVAE, a retrieval-augmented language model built upon the variational auto-encoder (VAE)
It encodes the text corpus into a latent space, capturing current and future information from both source and target text.
Experimental results on various datasets demonstrate significant improvements in text generation quality and hallucination removal.
arXiv Detail & Related papers (2023-10-16T16:42:01Z) - Boosting Event Extraction with Denoised Structure-to-Text Augmentation [52.21703002404442]
Event extraction aims to recognize pre-defined event triggers and arguments from texts.
Recent data augmentation methods often neglect the problem of grammatical incorrectness.
We propose a denoised structure-to-text augmentation framework for event extraction DAEE.
arXiv Detail & Related papers (2023-05-16T16:52:07Z) - Comparing Computational Architectures for Automated Journalism [0.0]
This study compares the most often employed methods for generating Brazilian Portuguese texts from structured data.
Results suggest that explicit intermediate steps in the generation process produce better texts than the ones generated by neural end-to-end architectures.
arXiv Detail & Related papers (2022-10-08T21:20:52Z) - Data-to-text Generation with Variational Sequential Planning [74.3955521225497]
We consider the task of data-to-text generation, which aims to create textual output from non-linguistic input.
We propose a neural model enhanced with a planning component responsible for organizing high-level information in a coherent and meaningful way.
We infer latent plans sequentially with a structured variational model, while interleaving the steps of planning and generation.
arXiv Detail & Related papers (2022-02-28T13:17:59Z) - Plan-then-Generate: Controlled Data-to-Text Generation via Planning [11.127156275580305]
We propose a novel Plan-then-Generate (PlanGen) framework to improve the controllability of neural data-to-text models.
Our model is able to control both the intra-sentence and inter-sentence structure of the generated output.
arXiv Detail & Related papers (2021-08-31T10:53:32Z) - SDA: Improving Text Generation with Self Data Augmentation [88.24594090105899]
We propose to improve the standard maximum likelihood estimation (MLE) paradigm by incorporating a self-imitation-learning phase for automatic data augmentation.
Unlike most existing sentence-level augmentation strategies, our method is more general and could be easily adapted to any MLE-based training procedure.
arXiv Detail & Related papers (2021-01-02T01:15:57Z) - Lightweight, Dynamic Graph Convolutional Networks for AMR-to-Text
Generation [56.73834525802723]
Lightweight Dynamic Graph Convolutional Networks (LDGCNs) are proposed.
LDGCNs capture richer non-local interactions by synthesizing higher order information from the input graphs.
We develop two novel parameter saving strategies based on the group graph convolutions and weight tied convolutions to reduce memory usage and model complexity.
arXiv Detail & Related papers (2020-10-09T06:03:46Z) - Schema-Guided Natural Language Generation [13.11874946084068]
We present the novel task ofGuided Natural Language Generation (SG-NLG)
In SG-NLG, the goal is still to generate a natural language prompt, but in SG-NLG, the input MRs are paired with rich schemata providing contextual information.
We train different state-of-the-art models for neural natural language generation on this dataset and show that in many cases, including rich schema information allows our models to produce higher quality outputs.
arXiv Detail & Related papers (2020-05-11T23:01:22Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z) - Have Your Text and Use It Too! End-to-End Neural Data-to-Text Generation
with Semantic Fidelity [3.8673630752805432]
We present DataTuner, a neural, end-to-end data-to-text generation system.
We take a two-stage generation-reranking approach, combining a fine-tuned language model with a semantic fidelity.
We show that DataTuner achieves state of the art results on the automated metrics across four major D2T datasets.
arXiv Detail & Related papers (2020-04-08T11:16:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.