Summary Grounded Conversation Generation
- URL: http://arxiv.org/abs/2106.03337v1
- Date: Mon, 7 Jun 2021 04:46:31 GMT
- Title: Summary Grounded Conversation Generation
- Authors: Chulaka Gunasekara, Guy Feigenblat, Benjamin Sznajder, Sachindra
Joshi, David Konopnicki
- Abstract summary: We show how pre-trained language models can be used to generate entire conversations, given only a summary of a conversation as the input.
We also show that the accuracy of conversation summarization can be improved by augmenting a conversation summarization dataset with generated conversations.
- Score: 10.470157142861174
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many conversation datasets have been constructed in the recent years using
crowdsourcing. However, the data collection process can be time consuming and
presents many challenges to ensure data quality. Since language generation has
improved immensely in recent years with the advancement of pre-trained language
models, we investigate how such models can be utilized to generate entire
conversations, given only a summary of a conversation as the input. We explore
three approaches to generate summary grounded conversations, and evaluate the
generated conversations using automatic measures and human judgements. We also
show that the accuracy of conversation summarization can be improved by
augmenting a conversation summarization dataset with generated conversations.
Related papers
- Evaluating Large Language Models in Semantic Parsing for Conversational
Question Answering over Knowledge Graphs [6.869834883252353]
This paper evaluates the performance of large language models that have not been explicitly pre-trained on this task.
Our results demonstrate that large language models are capable of generating graph queries from dialogues.
arXiv Detail & Related papers (2024-01-03T12:28:33Z) - Large Language Models as Zero-Shot Conversational Recommenders [52.57230221644014]
We present empirical studies on conversational recommendation tasks using representative large language models in a zero-shot setting.
We construct a new dataset of recommendation-related conversations by scraping a popular discussion website.
We observe that even without fine-tuning, large language models can outperform existing fine-tuned conversational recommendation models.
arXiv Detail & Related papers (2023-08-19T15:29:45Z) - AutoConv: Automatically Generating Information-seeking Conversations
with Large Language Models [74.10293412011455]
We propose AutoConv for synthetic conversation generation.
Specifically, we formulate the conversation generation problem as a language modeling task.
We finetune an LLM with a few human conversations to capture the characteristics of the information-seeking process.
arXiv Detail & Related papers (2023-08-12T08:52:40Z) - PLACES: Prompting Language Models for Social Conversation Synthesis [103.94325597273316]
We use a small set of expert-written conversations as in-context examples to synthesize a social conversation dataset using prompting.
We perform several thorough evaluations of our synthetic conversations compared to human-collected conversations.
arXiv Detail & Related papers (2023-02-07T05:48:16Z) - Talk the Walk: Synthetic Data Generation for Conversational Music
Recommendation [62.019437228000776]
We present TalkWalk, which generates realistic high-quality conversational data by leveraging encoded expertise in widely available item collections.
We generate over one million diverse conversations in a human-collected dataset.
arXiv Detail & Related papers (2023-01-27T01:54:16Z) - Knowledge-Grounded Conversational Data Augmentation with Generative
Conversational Networks [76.11480953550013]
We take a step towards automatically generating conversational data using Generative Conversational Networks.
We evaluate our approach on conversations with and without knowledge on the Topical Chat dataset.
arXiv Detail & Related papers (2022-07-22T22:37:14Z) - Training Conversational Agents with Generative Conversational Networks [74.9941330874663]
We use Generative Conversational Networks to automatically generate data and train social conversational agents.
We evaluate our approach on TopicalChat with automatic metrics and human evaluators, showing that with 10% of seed data it performs close to the baseline that uses 100% of the data.
arXiv Detail & Related papers (2021-10-15T21:46:39Z) - A Neural Conversation Generation Model via Equivalent Shared Memory
Investigation [39.922967513749654]
We propose a novel reading and memory framework called Deep Reading Memory Network (DRMN)
DRMN is capable of remembering useful information of similar conversations for improving utterance generation.
We apply our model to two large-scale conversation datasets of justice and e-commerce fields.
arXiv Detail & Related papers (2021-08-20T13:20:14Z) - Beyond Goldfish Memory: Long-Term Open-Domain Conversation [43.37382902468993]
We release a human-human dataset consisting of multiple chat sessions.
We show how existing models trained on existing datasets perform poorly in this long-term conversation setting.
In particular, we find retrieval-augmented methods and methods with an ability to summarize and recall previous conversations.
arXiv Detail & Related papers (2021-07-15T19:01:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.