Sketch and Refine: Towards Faithful and Informative Table-to-Text
Generation
- URL: http://arxiv.org/abs/2105.14778v1
- Date: Mon, 31 May 2021 08:18:13 GMT
- Title: Sketch and Refine: Towards Faithful and Informative Table-to-Text
Generation
- Authors: Peng Wang, Junyang Lin, An Yang, Chang Zhou, Yichang Zhang, Jingren
Zhou, Hongxia Yang
- Abstract summary: We propose a novel two-stage method that combines both Autoregressive and Non-Autoregressive generations (SANA)
Our approach includes: (1) skeleton generation with an autoregressive pointer network to select key tokens from the source table; (2) edit-based non-autoregressive generation model to produce texts via iterative insertion and deletion operations.
By integrating hard constraints from the skeleton, the non-autoregressive model improves the generation's coverage over the source table and thus enhances its faithfulness.
- Score: 58.320248632121476
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Table-to-text generation refers to generating a descriptive text from a
key-value table. Traditional autoregressive methods, though can generate text
with high fluency, suffer from low coverage and poor faithfulness problems. To
mitigate these problems, we propose a novel Skeleton-based two-stage method
that combines both Autoregressive and Non-Autoregressive generations (SANA).
Our approach includes: (1) skeleton generation with an autoregressive pointer
network to select key tokens from the source table; (2) edit-based
non-autoregressive generation model to produce texts via iterative insertion
and deletion operations. By integrating hard constraints from the skeleton, the
non-autoregressive model improves the generation's coverage over the source
table and thus enhances its faithfulness. We conduct automatic and human
evaluations on both WikiPerson and WikiBio datasets. Experimental results
demonstrate that our method outperforms the previous state-of-the-art methods
in both automatic and human evaluation, especially on coverage and
faithfulness. In particular, we achieve PARENT-T recall of 99.47 in WikiPerson,
improving over the existing best results by more than 10 points.
Related papers
- Enhancing Text Generation in Joint NLG/NLU Learning Through Curriculum Learning, Semi-Supervised Training, and Advanced Optimization Techniques [0.0]
This research paper developed a novel approach to improve text generation in the context of joint Natural Language Generation (NLG) and Natural Language Understanding (NLU) learning.
The data is prepared by gathering and preprocessing annotated datasets, including cleaning, tokenization, stemming, and stop-word removal.
Transformer-based encoders and decoders, capturing long range dependencies and improving source-target sequence modelling.
Reinforcement learning with policy gradient techniques, semi-supervised training, improved attention mechanisms, and differentiable approximations are employed to fine-tune the models and handle complex linguistic tasks effectively.
arXiv Detail & Related papers (2024-10-17T12:43:49Z) - Adversarial Counterfactual Augmentation: Application in Alzheimer's
Disease Classification [18.396331077506296]
We construct an adversarial game where we update the input textitconditional factor of the generator and the downstream textitclassifier with backpropagation.
We show the proposed approach improves test accuracy and can alleviate spurious correlations.
arXiv Detail & Related papers (2022-03-15T12:11:05Z) - Attend, Memorize and Generate: Towards Faithful Table-to-Text Generation
in Few Shots [58.404516361586325]
Few-shot table-to-text generation is a task of composing fluent and faithful sentences to convey table content using limited data.
This paper proposes a novel approach, Memorize and Generate (called AMG), inspired by the text generation process of humans.
arXiv Detail & Related papers (2022-03-01T20:37:20Z) - Facts2Story: Controlling Text Generation by Key Facts [0.0]
We propose a controlled generation task based on expanding a sequence of facts, expressed in natural language, into a longer narrative.
We show that while auto-regressive, unidirectional Language Models such as GPT2 produce better fluency, they struggle to adhere to the requested facts.
We propose a plan-and-cloze model (using fine-tuned XLNet) which produces competitive fluency while adhering to the requested content.
arXiv Detail & Related papers (2020-12-08T10:14:29Z) - Conditioned Text Generation with Transfer for Closed-Domain Dialogue
Systems [65.48663492703557]
We show how to optimally train and control the generation of intent-specific sentences using a conditional variational autoencoder.
We introduce a new protocol called query transfer that allows to leverage a large unlabelled dataset.
arXiv Detail & Related papers (2020-11-03T14:06:10Z) - Towards Faithful Neural Table-to-Text Generation with Content-Matching
Constraints [63.84063384518667]
We propose a novel Transformer-based generation framework to achieve the goal.
Core techniques in our method to enforce faithfulness include a new table-text optimal-transport matching loss.
To evaluate faithfulness, we propose a new automatic metric specialized to the table-to-text generation problem.
arXiv Detail & Related papers (2020-05-03T02:54:26Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z) - BERT-ATTACK: Adversarial Attack Against BERT Using BERT [77.82947768158132]
Adrial attacks for discrete data (such as texts) are more challenging than continuous data (such as images)
We propose textbfBERT-Attack, a high-quality and effective method to generate adversarial samples.
Our method outperforms state-of-the-art attack strategies in both success rate and perturb percentage.
arXiv Detail & Related papers (2020-04-21T13:30:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.