Text Editing as Imitation Game
- URL: http://arxiv.org/abs/2210.12276v1
- Date: Fri, 21 Oct 2022 22:07:04 GMT
- Title: Text Editing as Imitation Game
- Authors: Ning Shi, Bin Tang, Bo Yuan, Longtao Huang, Yewen Pu, Jie Fu, Zhouhan
Lin
- Abstract summary: We reformulate text editing as an imitation game using behavioral cloning.
We introduce a dual decoders structure to parallel the decoding while retaining the dependencies between action tokens.
Our model consistently outperforms the autoregressive baselines in terms of performance, efficiency, and robustness.
- Score: 33.418628166176234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text editing, such as grammatical error correction, arises naturally from
imperfect textual data. Recent works frame text editing as a multi-round
sequence tagging task, where operations -- such as insertion and substitution
-- are represented as a sequence of tags. While achieving good results, this
encoding is limited in flexibility as all actions are bound to token-level
tags. In this work, we reformulate text editing as an imitation game using
behavioral cloning. Specifically, we convert conventional sequence-to-sequence
data into state-to-action demonstrations, where the action space can be as
flexible as needed. Instead of generating the actions one at a time, we
introduce a dual decoders structure to parallel the decoding while retaining
the dependencies between action tokens, coupled with trajectory augmentation to
alleviate the distribution shift that imitation learning often suffers. In
experiments on a suite of Arithmetic Equation benchmarks, our model
consistently outperforms the autoregressive baselines in terms of performance,
efficiency, and robustness. We hope our findings will shed light on future
studies in reinforcement learning applying sequence-level action generation to
natural language processing.
Related papers
- Text2Data: Low-Resource Data Generation with Textual Control [104.38011760992637]
Natural language serves as a common and straightforward control signal for humans to interact seamlessly with machines.
We propose Text2Data, a novel approach that utilizes unlabeled data to understand the underlying data distribution through an unsupervised diffusion model.
It undergoes controllable finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - Co-Speech Gesture Detection through Multi-Phase Sequence Labeling [3.924524252255593]
We introduce a novel framework that reframes the task as a multi-phase sequence labeling problem.
We evaluate our proposal on a large dataset of diverse co-speech gestures in task-oriented face-to-face dialogues.
arXiv Detail & Related papers (2023-08-21T12:27:18Z) - Real-World Compositional Generalization with Disentangled
Sequence-to-Sequence Learning [81.24269148865555]
A recently proposed Disentangled sequence-to-sequence model (Dangle) shows promising generalization capability.
We introduce two key modifications to this model which encourage more disentangled representations and improve its compute and memory efficiency.
Specifically, instead of adaptively re-encoding source keys and values at each time step, we disentangle their representations and only re-encode keys periodically.
arXiv Detail & Related papers (2022-12-12T15:40:30Z) - Composable Text Controls in Latent Space with ODEs [97.12426987887021]
This paper proposes a new efficient approach for composable text operations in the compact latent space of text.
By connecting pretrained LMs to the latent space through efficient adaption, we then decode the sampled vectors into desired text sequences.
Experiments show that composing those operators within our approach manages to generate or edit high-quality text.
arXiv Detail & Related papers (2022-08-01T06:51:45Z) - Text Revision by On-the-Fly Representation Optimization [76.11035270753757]
Current state-of-the-art methods formulate these tasks as sequence-to-sequence learning problems.
We present an iterative in-place editing approach for text revision, which requires no parallel data.
It achieves competitive and even better performance than state-of-the-art supervised methods on text simplification.
arXiv Detail & Related papers (2022-04-15T07:38:08Z) - Discontinuous Grammar as a Foreign Language [0.7412445894287709]
We extend the framework of sequence-to-sequence models for constituent parsing.
We design several novelizations that can fully produce discontinuities.
For the first time, we test a sequence-to-sequence model on the main discontinuous benchmarks.
arXiv Detail & Related papers (2021-10-20T08:58:02Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - Seq2Edits: Sequence Transduction Using Span-level Edit Operations [10.785577504399077]
Seq2Edits is an open-vocabulary approach to sequence editing for natural language processing (NLP) tasks.
We evaluate our method on five NLP tasks (text normalization, sentence fusion, sentence splitting & rephrasing, text simplification, and grammatical error correction)
For grammatical error correction, our method speeds up inference by up to 5.2x compared to full sequence models.
arXiv Detail & Related papers (2020-09-23T13:28:38Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.