MultiDoc2Dial: Modeling Dialogues Grounded in Multiple Documents
- URL: http://arxiv.org/abs/2109.12595v1
- Date: Sun, 26 Sep 2021 13:12:05 GMT
- Title: MultiDoc2Dial: Modeling Dialogues Grounded in Multiple Documents
- Authors: Song Feng and Siva Sankalp Patel and Hui Wan and Sachindra Joshi
- Abstract summary: We propose MultiDoc2Dial, a new task and dataset on modeling goal-oriented dialogues grounded in multiple documents.
We introduce a new dataset that contains dialogues grounded in multiple documents from four different domains.
- Score: 14.807409907211452
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose MultiDoc2Dial, a new task and dataset on modeling goal-oriented
dialogues grounded in multiple documents. Most previous works treat
document-grounded dialogue modeling as a machine reading comprehension task
based on a single given document or passage. In this work, we aim to address
more realistic scenarios where a goal-oriented information-seeking conversation
involves multiple topics, and hence is grounded on different documents. To
facilitate such a task, we introduce a new dataset that contains dialogues
grounded in multiple documents from four different domains. We also explore
modeling the dialogue-based and document-based context in the dataset. We
present strong baseline approaches and various experimental results, aiming to
support further research efforts on such a task.
Related papers
- Unified Multi-Modal Interleaved Document Representation for Information Retrieval [57.65409208879344]
We produce more comprehensive and nuanced document representations by holistically embedding documents interleaved with different modalities.
Specifically, we achieve this by leveraging the capability of recent vision-language models that enable the processing and integration of text, images, and tables into a unified format and representation.
arXiv Detail & Related papers (2024-10-03T17:49:09Z) - DialogStudio: Towards Richest and Most Diverse Unified Dataset
Collection for Conversational AI [92.29874802394167]
DialogStudio is the largest and most diverse collection of dialogue datasets.
Our collection encompasses data from open-domain dialogues, task-oriented dialogues, natural language understanding, conversational recommendation, dialogue summarization, and knowledge-grounded dialogues.
arXiv Detail & Related papers (2023-07-19T17:57:53Z) - Doc2Bot: Accessing Heterogeneous Documents via Conversational Bots [103.54897676954091]
Doc2Bot is a dataset for building machines that help users seek information via conversations.
Our dataset contains over 100,000 turns based on Chinese documents from five domains.
arXiv Detail & Related papers (2022-10-20T07:33:05Z) - Manual-Guided Dialogue for Flexible Conversational Agents [84.46598430403886]
How to build and use dialogue data efficiently, and how to deploy models in different domains at scale can be critical issues in building a task-oriented dialogue system.
We propose a novel manual-guided dialogue scheme, where the agent learns the tasks from both dialogue and manuals.
Our proposed scheme reduces the dependence of dialogue models on fine-grained domain ontology, and makes them more flexible to adapt to various domains.
arXiv Detail & Related papers (2022-08-16T08:21:12Z) - DialogZoo: Large-Scale Dialog-Oriented Task Learning [52.18193690394549]
We aim to build a unified foundation model which can solve massive diverse dialogue tasks.
To achieve this goal, we first collect a large-scale well-labeled dialogue dataset from 73 publicly available datasets.
arXiv Detail & Related papers (2022-05-25T11:17:16Z) - Constructing Multi-Modal Dialogue Dataset by Replacing Text with
Semantically Relevant Images [17.076424447172297]
This paper proposes a 45k multi-modal dialogue dataset created with minimal human intervention.
Our method to create such a dataset consists of (1) preparing and pre-processing text dialogue datasets, (2) creating image-mixed dialogues by using a text-to-image replacement technique, and (3) employing a contextual-similarity-based filtering step.
arXiv Detail & Related papers (2021-07-19T08:44:11Z) - Converse, Focus and Guess -- Towards Multi-Document Driven Dialogue [53.380996227212165]
We propose a novel task, Multi-Document Driven Dialogue (MD3), in which an agent can guess the target document that the user is interested in by leading a dialogue.
GuessMovie contains 16,881 documents, each describing a movie, and associated 13,434 dialogues.
Our method significantly outperforms several strong baseline methods and is very close to human's performance.
arXiv Detail & Related papers (2021-02-04T06:36:11Z) - doc2dial: A Goal-Oriented Document-Grounded Dialogue Dataset [24.040517978408484]
doc2dial is a new dataset of goal-oriented dialogues grounded in documents.
We first construct dialogue flows based on the content elements that corresponds to higher-level relations across text sections.
We present these dialogue flows to crowd contributors to create conversational utterances.
arXiv Detail & Related papers (2020-11-12T19:08:44Z) - Towards a Multi-modal, Multi-task Learning based Pre-training Framework
for Document Representation Learning [5.109216329453963]
We introduce Document Topic Modelling and Document Shuffle Prediction as novel pre-training tasks.
We utilize the Longformer network architecture as the backbone to encode the multi-modal information from multi-page documents in an end-to-end fashion.
arXiv Detail & Related papers (2020-09-30T05:39:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.