Turning Flowchart into Dialog: Augmenting Flowchart-grounded
Troubleshooting Dialogs via Synthetic Data Generation
- URL: http://arxiv.org/abs/2305.01323v3
- Date: Sun, 29 Oct 2023 11:02:47 GMT
- Title: Turning Flowchart into Dialog: Augmenting Flowchart-grounded
Troubleshooting Dialogs via Synthetic Data Generation
- Authors: Haolan Zhan and Sameen Maruf and Lizhen Qu and Yufei Wang and Ingrid
Zukerman and Gholamreza Haffari
- Abstract summary: Flowchart-grounded troubleshooting dialogue (FTD) systems follow the instructions of a flowchart to diagnose users' problems in specific domains.
We propose a plan-based synthetic data generation approach that generates diverse synthetic dialog data at scale.
- Score: 50.06143883455979
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Flowchart-grounded troubleshooting dialogue (FTD) systems, which follow the
instructions of a flowchart to diagnose users' problems in specific domains
(e.g., vehicle, laptop), have been gaining research interest in recent years.
However, collecting sufficient dialogues that are naturally grounded on
flowcharts is costly, thus FTD systems are impeded by scarce training data. To
mitigate the data sparsity issue, we propose a plan-based synthetic data
generation (PlanSDG) approach that generates diverse synthetic dialog data at
scale by transforming concise flowchart into dialogues. Specifically, its
generative model employs a variational-base framework with a hierarchical
planning strategy that includes global and local latent planning variables.
Experiments on the FloDial dataset show that synthetic dialogue produced by
PlanSDG improves the performance of downstream tasks, including flowchart path
retrieval and response generation, in particular on the Out-of-Flowchart
settings. In addition, further analysis demonstrate the quality of synthetic
data generated by PlanSDG in paths that are covered by current sample dialogues
and paths that are not covered.
Related papers
- ToolFlow: Boosting LLM Tool-Calling Through Natural and Coherent Dialogue Synthesis [80.34000499166648]
We propose a Graph-based Sampling strategy to sample more relevant tool combinations, and a Planned-generation strategy to create plans that guide the synthesis of coherent dialogues.
We apply SFT on LLaMA-3.1-8B using 8,000 synthetic dialogues generated with ToolFlow.
Results show that the model achieves tool-calling performance comparable to or even surpassing GPT-4, while maintaining strong general capabilities.
arXiv Detail & Related papers (2024-10-24T05:45:04Z) - DFlow: Diverse Dialogue Flow Simulation with Large Language Models [16.209331014315463]
This paper proposes a novel data augmentation method designed to enhance the diversity of synthetic dialogues.
We generate a task-oriented dialogue dataset comprising 3,886 dialogue flows across 15 different domains.
arXiv Detail & Related papers (2024-10-18T20:35:28Z) - Simulating Task-Oriented Dialogues with State Transition Graphs and Large Language Models [16.94819621353007]
SynTOD is a new synthetic data generation approach for developing end-to-end Task-Oriented Dialogue (TOD) systems.
It generates diverse, structured conversations through random walks and response simulation using large language models.
In our experiments, using graph-guided response simulations leads to significant improvements in intent classification, slot filling and response relevance.
arXiv Detail & Related papers (2024-04-23T06:23:34Z) - TOD-Flow: Modeling the Structure of Task-Oriented Dialogues [77.15457469745364]
We propose a novel approach focusing on inferring the TOD-Flow graph from dialogue data annotated with dialog acts.
The inferred TOD-Flow graph can be easily integrated with any dialogue model to improve its prediction performance, transparency, and controllability.
arXiv Detail & Related papers (2023-12-07T20:06:23Z) - Local-Global Information Interaction Debiasing for Dynamic Scene Graph
Generation [51.92419880088668]
We propose a novel DynSGG model based on multi-task learning, DynSGG-MTL, which introduces the local interaction information and global human-action interaction information.
Long-temporal human actions supervise the model to generate multiple scene graphs that conform to the global constraints and avoid the model being unable to learn the tail predicates.
arXiv Detail & Related papers (2023-08-10T01:24:25Z) - End-to-End Learning of Flowchart Grounded Task-Oriented Dialogs [23.678209058054062]
We propose a novel problem within end-to-end learning of task-oriented dialogs (TOD)
Such dialogs are grounded in domain-specific flowcharts, which the agent is supposed to follow during the conversation.
We release a dataset consisting of 2,738 dialogs grounded on 12 different troubleshooting flowcharts.
We also design a neural model, FloNet, which uses a retrieval-augmented generation architecture to train the dialog agent.
arXiv Detail & Related papers (2021-09-15T12:58:51Z) - Task-Oriented Dialogue as Dataflow Synthesis [158.77123205487334]
We describe an approach to task-oriented dialogue in which dialogue state is represented as a dataflow graph.
A dialogue agent maps each user utterance to a program that extends this graph.
We introduce a new dataset, SMCalFlow, featuring complex dialogues about events, weather, places, and people.
arXiv Detail & Related papers (2020-09-24T00:35:26Z) - Variational Hierarchical Dialog Autoencoder for Dialog State Tracking
Data Augmentation [59.174903564894954]
In this work, we extend this approach to the task of dialog state tracking for goal-oriented dialogs.
We propose the Variational Hierarchical Dialog Autoencoder (VHDA) for modeling the complete aspects of goal-oriented dialogs.
Experiments on various dialog datasets show that our model improves the downstream dialog trackers' robustness via generative data augmentation.
arXiv Detail & Related papers (2020-01-23T15:34:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.