Gated Mechanism Enhanced Multi-Task Learning for Dialog Routing
- URL: http://arxiv.org/abs/2304.03730v1
- Date: Fri, 7 Apr 2023 16:51:46 GMT
- Title: Gated Mechanism Enhanced Multi-Task Learning for Dialog Routing
- Authors: Ziming Huang and Zhuoxuan Jiang and Ke Wang and Juntao Li and Shanshan
Feng and Xian-Ling Mao
- Abstract summary: Gated Mechanism enhanced Multi-task Model (G3M)
Proposal includes a novel dialog encoder and two tailored gated mechanism modules.
Based on two datasets collected from real world applications, extensive experimental results demonstrate the effectiveness of our method.
- Score: 28.870359916550996
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Currently, human-bot symbiosis dialog systems, e.g., pre- and after-sales in
E-commerce, are ubiquitous, and the dialog routing component is essential to
improve the overall efficiency, reduce human resource cost, and enhance user
experience. Although most existing methods can fulfil this requirement, they
can only model single-source dialog data and cannot effectively capture the
underlying knowledge of relations among data and subtasks. In this paper, we
investigate this important problem by thoroughly mining both the data-to-task
and task-to-task knowledge among various kinds of dialog data. To achieve the
above targets, we propose a Gated Mechanism enhanced Multi-task Model (G3M),
specifically including a novel dialog encoder and two tailored gated mechanism
modules. The proposed method can play the role of hierarchical information
filtering and is non-invasive to existing dialog systems. Based on two datasets
collected from real world applications, extensive experimental results
demonstrate the effectiveness of our method, which achieves the
state-of-the-art performance by improving 8.7\%/11.8\% on RMSE metric and
2.2\%/4.4\% on F1 metric.
Related papers
- DialCLIP: Empowering CLIP as Multi-Modal Dialog Retriever [83.33209603041013]
We propose a parameter-efficient prompt-tuning method named DialCLIP for multi-modal dialog retrieval.
Our approach introduces a multi-modal context generator to learn context features which are distilled into prompts within the pre-trained vision-language model CLIP.
To facilitate various types of retrieval, we also design multiple experts to learn mappings from CLIP outputs to multi-modal representation space.
arXiv Detail & Related papers (2024-01-02T07:40:12Z) - Application of frozen large-scale models to multimodal task-oriented
dialogue [0.0]
We use the existing Large Language Models ENnhanced to See Framework (LENS Framework) to test the feasibility of multimodal task-oriented dialogues.
The LENS Framework has been proposed as a method to solve computer vision tasks without additional training and with fixed parameters of pre-trained models.
arXiv Detail & Related papers (2023-10-02T01:42:28Z) - Weakly Supervised Data Augmentation Through Prompting for Dialogue
Understanding [103.94325597273316]
We present a novel approach that iterates on augmentation quality by applying weakly-supervised filters.
We evaluate our methods on the emotion and act classification tasks in DailyDialog and the intent classification task in Facebook Multilingual Task-Oriented Dialogue.
For DailyDialog specifically, using 10% of the ground truth data we outperform the current state-of-the-art model which uses 100% of the data.
arXiv Detail & Related papers (2022-10-25T17:01:30Z) - Towards End-to-End Open Conversational Machine Reading [57.18251784418258]
In open-retrieval conversational machine reading (OR-CMR) task, machines are required to do multi-turn question answering given dialogue history and a textual knowledge base.
We model OR-CMR as a unified text-to-text task in a fully end-to-end style. Experiments on the ShARC and OR-ShARC dataset show the effectiveness of our proposed end-to-end framework.
arXiv Detail & Related papers (2022-10-13T15:50:44Z) - Manual-Guided Dialogue for Flexible Conversational Agents [84.46598430403886]
How to build and use dialogue data efficiently, and how to deploy models in different domains at scale can be critical issues in building a task-oriented dialogue system.
We propose a novel manual-guided dialogue scheme, where the agent learns the tasks from both dialogue and manuals.
Our proposed scheme reduces the dependence of dialogue models on fine-grained domain ontology, and makes them more flexible to adapt to various domains.
arXiv Detail & Related papers (2022-08-16T08:21:12Z) - KETOD: Knowledge-Enriched Task-Oriented Dialogue [77.59814785157877]
Existing studies in dialogue system research mostly treat task-oriented dialogue and chit-chat as separate domains.
We investigate how task-oriented dialogue and knowledge-grounded chit-chat can be effectively integrated into a single model.
arXiv Detail & Related papers (2022-05-11T16:01:03Z) - Quick Starting Dialog Systems with Paraphrase Generation [0.0]
We propose a method to reduce the cost and effort of creating new conversational agents by artificially generating more data from existing examples.
Our proposed approach can kick-start a dialog system with little human effort, and brings its performance to a level satisfactory enough for allowing actual interactions with real end-users.
arXiv Detail & Related papers (2022-04-06T02:35:59Z) - Data-Efficient Methods for Dialogue Systems [4.061135251278187]
Conversational User Interface (CUI) has become ubiquitous in everyday life, in consumer-focused products like Siri and Alexa.
Deep learning underlies many recent breakthroughs in dialogue systems but requires very large amounts of training data, often annotated by experts.
In this thesis, we introduce a series of methods for training robust dialogue systems from minimal data.
arXiv Detail & Related papers (2020-12-05T02:51:09Z) - DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification [77.59549450705384]
In dialog system, dialog act recognition and sentiment classification are two correlative tasks.
Most of the existing systems either treat them as separate tasks or just jointly model the two tasks.
We propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly consider the cross-impact and model the interaction between the two tasks.
arXiv Detail & Related papers (2020-08-16T14:13:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.