RSVP: Customer Intent Detection via Agent Response Contrastive and
Generative Pre-Training
- URL: http://arxiv.org/abs/2310.09773v1
- Date: Sun, 15 Oct 2023 08:21:38 GMT
- Title: RSVP: Customer Intent Detection via Agent Response Contrastive and
Generative Pre-Training
- Authors: Yu-Chien Tang, Wei-Yao Wang, An-Zi Yen, Wen-Chih Peng
- Abstract summary: RSVP is a self-supervised framework dedicated to task-oriented dialogues.
It incorporates agent responses for pre-training in a two-stage manner.
Our benchmark results for two real-world customer service datasets show that RSVP significantly outperforms the state-of-the-art baselines.
- Score: 16.183398769581075
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The dialogue systems in customer services have been developed with neural
models to provide users with precise answers and round-the-clock support in
task-oriented conversations by detecting customer intents based on their
utterances. Existing intent detection approaches have highly relied on
adaptively pre-training language models with large-scale datasets, yet the
predominant cost of data collection may hinder their superiority. In addition,
they neglect the information within the conversational responses of the agents,
which have a lower collection cost, but are significant to customer intent as
agents must tailor their replies based on the customers' intent. In this paper,
we propose RSVP, a self-supervised framework dedicated to task-oriented
dialogues, which utilizes agent responses for pre-training in a two-stage
manner. Specifically, we introduce two pre-training tasks to incorporate the
relations of utterance-response pairs: 1) Response Retrieval by selecting a
correct response from a batch of candidates, and 2) Response Generation by
mimicking agents to generate the response to a given utterance. Our benchmark
results for two real-world customer service datasets show that RSVP
significantly outperforms the state-of-the-art baselines by 4.95% for accuracy,
3.4% for MRR@3, and 2.75% for MRR@5 on average. Extensive case studies are
investigated to show the validity of incorporating agent responses into the
pre-training stage.
Related papers
- RAG based Question-Answering for Contextual Response Prediction System [0.4660328753262075]
Large Language Models (LLMs) have shown versatility in various Natural Language Processing (NLP) tasks.
Retrieval Augmented Generation (RAG) emerges as a promising technique to address this challenge.
This paper introduces an end-to-end framework that employs LLMs with RAG capabilities for industry use cases.
arXiv Detail & Related papers (2024-09-05T17:14:23Z) - Selecting Query-bag as Pseudo Relevance Feedback for Information-seeking Conversations [76.70349332096693]
Information-seeking dialogue systems are widely used in e-commerce systems.
We propose a Query-bag based Pseudo Relevance Feedback framework (QB-PRF)
It constructs a query-bag with related queries to serve as pseudo signals to guide information-seeking conversations.
arXiv Detail & Related papers (2024-03-22T08:10:32Z) - Towards Reliable and Factual Response Generation: Detecting Unanswerable
Questions in Information-Seeking Conversations [16.99952884041096]
Generative AI models face the challenge of hallucinations that can undermine users' trust in such systems.
We approach the problem of conversational information seeking as a two-step process, where relevant passages in a corpus are identified first and then summarized into a final system response.
Specifically, our proposed method employs a sentence-level classifier to detect if the answer is present, then aggregates these predictions on the passage level, and eventually across the top-ranked passages to arrive at a final answerability estimate.
arXiv Detail & Related papers (2024-01-21T10:15:36Z) - Socratic Pretraining: Question-Driven Pretraining for Controllable
Summarization [89.04537372465612]
Socratic pretraining is a question-driven, unsupervised pretraining objective designed to improve controllability in summarization tasks.
Our results show that Socratic pretraining cuts task-specific labeled data requirements in half.
arXiv Detail & Related papers (2022-12-20T17:27:10Z) - A Systematic Evaluation of Response Selection for Open Domain Dialogue [36.88551817451512]
We curated a dataset where responses from multiple response generators produced for the same dialog context are manually annotated as appropriate (positive) and inappropriate (negative)
We conduct a systematic evaluation of state-of-the-art methods for response selection, and demonstrate that both strategies of using multiple positive candidates and using manually verified hard negative candidates can bring in significant performance improvement in comparison to using the adversarial training data, e.g., increase of 3% and 13% in Recall@1 score, respectively.
arXiv Detail & Related papers (2022-08-08T19:33:30Z) - Double Retrieval and Ranking for Accurate Question Answering [120.69820139008138]
We show that an answer verification step introduced in Transformer-based answer selection models can significantly improve the state of the art in Question Answering.
The results on three well-known datasets for AS2 show consistent and significant improvement of the state of the art.
arXiv Detail & Related papers (2022-01-16T06:20:07Z) - Partner Matters! An Empirical Study on Fusing Personas for Personalized
Response Selection in Retrieval-Based Chatbots [51.091235903442715]
This paper makes an attempt to explore the impact of utilizing personas that describe either self or partner speakers on the task of response selection.
Four persona fusion strategies are designed, which assume personas interact with contexts or responses in different ways.
Empirical studies on the Persona-Chat dataset show that the partner personas can improve the accuracy of response selection.
arXiv Detail & Related papers (2021-05-19T10:32:30Z) - Learning an Effective Context-Response Matching Model with
Self-Supervised Tasks for Retrieval-based Dialogues [88.73739515457116]
We introduce four self-supervised tasks including next session prediction, utterance restoration, incoherence detection and consistency discrimination.
We jointly train the PLM-based response selection model with these auxiliary tasks in a multi-task manner.
Experiment results indicate that the proposed auxiliary self-supervised tasks bring significant improvement for multi-turn response selection.
arXiv Detail & Related papers (2020-09-14T08:44:46Z) - Multi-Stage Conversational Passage Retrieval: An Approach to Fusing Term
Importance Estimation and Neural Query Rewriting [56.268862325167575]
We tackle conversational passage retrieval (ConvPR) with query reformulation integrated into a multi-stage ad-hoc IR system.
We propose two conversational query reformulation (CQR) methods: (1) term importance estimation and (2) neural query rewriting.
For the former, we expand conversational queries using important terms extracted from the conversational context with frequency-based signals.
For the latter, we reformulate conversational queries into natural, standalone, human-understandable queries with a pretrained sequence-tosequence model.
arXiv Detail & Related papers (2020-05-05T14:30:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.