Robust Dialogue Utterance Rewriting as Sequence Tagging
- URL: http://arxiv.org/abs/2012.14535v1
- Date: Tue, 29 Dec 2020 00:05:35 GMT
- Title: Robust Dialogue Utterance Rewriting as Sequence Tagging
- Authors: Jie Hao, Linfeng Song, Liwei Wang, Kun Xu, Zhaopeng Tu and Dong Yu
- Abstract summary: The task of dialogue rewriting aims to reconstruct the latest dialogue utterance by copying the missing content from the dialogue context.
Until now, the existing models for this task suffer from the robustness issue, i.e., performances drop dramatically when testing on a different domain.
We propose a novel sequence-tagging-based fluency model so that the search space is significantly reduced.
- Score: 62.12912805378693
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The task of dialogue rewriting aims to reconstruct the latest dialogue
utterance by copying the missing content from the dialogue context. Until now,
the existing models for this task suffer from the robustness issue, i.e.,
performances drop dramatically when testing on a different domain. We address
this robustness issue by proposing a novel sequence-tagging-based model so that
the search space is significantly reduced, yet the core of this task is still
well covered. As a common issue of most tagging models for text generation, the
model's outputs may lack fluency. To alleviate this issue, we inject the loss
signal from BLEU or GPT-2 under a REINFORCE framework. Experiments show huge
improvements of our model over the current state-of-the-art systems on domain
transfer.
Related papers
- Detecting Document-level Paraphrased Machine Generated Content: Mimicking Human Writing Style and Involving Discourse Features [57.34477506004105]
Machine-generated content poses challenges such as academic plagiarism and the spread of misinformation.
We introduce novel methodologies and datasets to overcome these challenges.
We propose MhBART, an encoder-decoder model designed to emulate human writing style.
We also propose DTransformer, a model that integrates discourse analysis through PDTB preprocessing to encode structural features.
arXiv Detail & Related papers (2024-12-17T08:47:41Z) - Cohesive Conversations: Enhancing Authenticity in Multi-Agent Simulated Dialogues [17.38671584773247]
This paper investigates the quality of multi-agent dialogues in simulations powered by Large Language Models (LLMs)
We propose a novel Screening, Diagnosis, and Regeneration (SDR) framework that detects and corrects utterance errors.
arXiv Detail & Related papers (2024-07-13T14:24:45Z) - Separate-and-Enhance: Compositional Finetuning for Text2Image Diffusion
Models [58.46926334842161]
This work illuminates the fundamental reasons for such misalignment, pinpointing issues related to low attention activation scores and mask overlaps.
We propose two novel objectives, the Separate loss and the Enhance loss, that reduce object mask overlaps and maximize attention scores.
Our method diverges from conventional test-time-adaptation techniques, focusing on finetuning critical parameters, which enhances scalability and generalizability.
arXiv Detail & Related papers (2023-12-10T22:07:42Z) - Incomplete Utterance Rewriting as Sequential Greedy Tagging [0.0]
We introduce speaker-aware embedding to model speaker variation.
Our model achieves optimal results on all nine restoration scores while having other metric scores comparable to previous state-of-the-art models.
arXiv Detail & Related papers (2023-07-08T04:05:04Z) - An Empirical Study of Multitask Learning to Improve Open Domain Dialogue
Systems [0.13706331473063876]
This paper describes an investigation where four different auxiliary tasks are added to small and medium-sized GPT-2 models.
The results show that the introduction of the new auxiliary tasks leads to small but consistent improvement in evaluations of the investigated models.
arXiv Detail & Related papers (2023-04-17T09:44:56Z) - Stabilized In-Context Learning with Pre-trained Language Models for Few
Shot Dialogue State Tracking [57.92608483099916]
Large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks.
For more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial.
We introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query.
arXiv Detail & Related papers (2023-02-12T15:05:10Z) - He Said, She Said: Style Transfer for Shifting the Perspective of
Dialogues [75.58367095888914]
We define a new style transfer task: perspective shift, which reframes a dialogue from informal first person to a formal third person rephrasing of the text.
As a sample application, we demonstrate that applying perspective shifting to a dialogue summarization dataset (SAMSum) substantially improves the zero-shot performance of extractive news summarization models.
arXiv Detail & Related papers (2022-10-27T14:16:07Z) - DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for
Dialog Response Generation [80.45816053153722]
DialogVED introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses.
We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation.
arXiv Detail & Related papers (2022-04-27T16:18:15Z) - Task-Oriented Dialogue System as Natural Language Generation [29.870260635814436]
We propose to formulate the task-oriented dialogue system as the purely natural language generation task.
This method heavily suffers from the dialogue entity inconsistency caused by the removal of delexicalized tokens.
We design a novel GPT-Adapter-CopyNet network, which incorporates the lightweight adapter and CopyNet modules into GPT-2 to achieve better performance.
arXiv Detail & Related papers (2021-08-31T08:36:42Z) - SARG: A Novel Semi Autoregressive Generator for Multi-turn Incomplete
Utterance Restoration [9.394277095571942]
In this paper, we investigate the incomplete utterance restoration which has brought general improvement over multi-turn dialogue systems.
We propose a novel semi autoregressive generator (SARG) with the high efficiency and flexibility.
arXiv Detail & Related papers (2020-08-04T11:52:20Z) - PALM: Pre-training an Autoencoding&Autoregressive Language Model for
Context-conditioned Generation [92.7366819044397]
Self-supervised pre-training has emerged as a powerful technique for natural language understanding and generation.
This work presents PALM with a novel scheme that jointly pre-trains an autoencoding and autoregressive language model on a large unlabeled corpus.
An extensive set of experiments show that PALM achieves new state-of-the-art results on a variety of language generation benchmarks.
arXiv Detail & Related papers (2020-04-14T06:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.