A Unified Multi-task Learning Framework for Multi-goal Conversational
Recommender Systems
- URL: http://arxiv.org/abs/2204.06923v1
- Date: Thu, 14 Apr 2022 12:31:27 GMT
- Title: A Unified Multi-task Learning Framework for Multi-goal Conversational
Recommender Systems
- Authors: Yang Deng, Wenxuan Zhang, Weiwen Xu, Wenqiang Lei, Tat-Seng Chua, Wai
Lam
- Abstract summary: Four tasks are often involved in MG-CRS, including Goal Planning, Topic Prediction, Item Recommendation, and Response Generation.
We propose a novel Unified MultI-goal conversational recommeNDer system, namely UniMIND.
Prompt-based learning strategies are investigated to endow the unified model with the capability of multi-task learning.
- Score: 91.70511776167488
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent years witnessed several advances in developing multi-goal
conversational recommender systems (MG-CRS) that can proactively attract users'
interests and naturally lead user-engaged dialogues with multiple
conversational goals and diverse topics. Four tasks are often involved in
MG-CRS, including Goal Planning, Topic Prediction, Item Recommendation, and
Response Generation. Most existing studies address only some of these tasks. To
handle the whole problem of MG-CRS, modularized frameworks are adopted where
each task is tackled independently without considering their interdependencies.
In this work, we propose a novel Unified MultI-goal conversational recommeNDer
system, namely UniMIND. In specific, we unify these four tasks with different
formulations into the same sequence-to-sequence (Seq2Seq) paradigm.
Prompt-based learning strategies are investigated to endow the unified model
with the capability of multi-task learning. Finally, the overall learning and
inference procedure consists of three stages, including multi-task learning,
prompt-based tuning, and inference. Experimental results on two MG-CRS
benchmarks (DuRecDial and TG-ReDial) show that UniMIND achieves
state-of-the-art performance on all tasks with a unified model. Extensive
analyses and discussions are provided for shedding some new perspectives for
MG-CRS.
Related papers
- MOMAland: A Set of Benchmarks for Multi-Objective Multi-Agent Reinforcement Learning [7.822825134714791]
Multi-objective multi-agent reinforcement learning (MOMARL) addresses problems with multiple agents each needing to consider multiple objectives in their learning process.
MOAland is the first collection of standardised environments for multi-objective multi-agent reinforcement learning.
arXiv Detail & Related papers (2024-07-23T09:05:06Z) - UniMS-RAG: A Unified Multi-source Retrieval-Augmented Generation for
Personalized Dialogue Systems [44.893215129952395]
Large Language Models (LLMs) has shown exceptional capabilities in many natual language understanding and generation tasks.
We decompose the use of multiple sources in generating personalized response into three sub-tasks: Knowledge Source Selection, Knowledge Retrieval, and Response Generation.
We propose a novel Unified Multi-Source Retrieval-Augmented Generation system (UniMS-RAG)
arXiv Detail & Related papers (2024-01-24T06:50:20Z) - MmAP : Multi-modal Alignment Prompt for Cross-domain Multi-task Learning [29.88567810099265]
Multi-task learning is designed to train multiple correlated tasks simultaneously.
To tackle this challenge, we integrate the decoder-free vision-language model CLIP.
We propose Multi-modal Alignment Prompt (MmAP) for CLIP, which aligns text and visual modalities during fine-tuning process.
arXiv Detail & Related papers (2023-12-14T03:33:02Z) - Multitask Multimodal Prompted Training for Interactive Embodied Task
Completion [48.69347134411864]
Embodied MultiModal Agent (EMMA) is a unified encoder-decoder model that reasons over images and trajectories.
By unifying all tasks as text generation, EMMA learns a language of actions which facilitates transfer across tasks.
arXiv Detail & Related papers (2023-11-07T15:27:52Z) - Revisiting Disentanglement and Fusion on Modality and Context in
Conversational Multimodal Emotion Recognition [81.2011058113579]
We argue that both the feature multimodality and conversational contextualization should be properly modeled simultaneously during the feature disentanglement and fusion steps.
We propose a Contribution-aware Fusion Mechanism (CFM) and a Context Refusion Mechanism ( CRM) for multimodal and context integration.
Our system achieves new state-of-the-art performance consistently.
arXiv Detail & Related papers (2023-08-08T18:11:27Z) - UniDU: Towards A Unified Generative Dialogue Understanding Framework [62.8474841241855]
We investigate a unified generative dialogue understanding framework, namely UniDU, to achieve information exchange among DU tasks.
We conduct experiments on ten dialogue understanding datasets, which span five fundamental tasks.
The proposed UniDU framework outperforms task-specific well-designed methods on all 5 tasks.
arXiv Detail & Related papers (2022-04-10T09:32:34Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - Masking Orchestration: Multi-task Pretraining for Multi-role Dialogue
Representation Learning [50.5572111079898]
Multi-role dialogue understanding comprises a wide range of diverse tasks such as question answering, act classification, dialogue summarization etc.
While dialogue corpora are abundantly available, labeled data, for specific learning tasks, can be highly scarce and expensive.
In this work, we investigate dialogue context representation learning with various types unsupervised pretraining tasks.
arXiv Detail & Related papers (2020-02-27T04:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.