Exploiting Unsupervised Data for Emotion Recognition in Conversations
- URL: http://arxiv.org/abs/2010.01908v2
- Date: Tue, 6 Oct 2020 09:32:31 GMT
- Title: Exploiting Unsupervised Data for Emotion Recognition in Conversations
- Authors: Wenxiang Jiao, Michael R. Lyu, Irwin King
- Abstract summary: Emotion Recognition in Conversations (ERC) aims to predict the emotional state of speakers in conversations.
The available supervised data for the ERC task is limited.
We propose a novel approach to leverage unsupervised conversation data.
- Score: 76.01690906995286
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Emotion Recognition in Conversations (ERC) aims to predict the emotional
state of speakers in conversations, which is essentially a text classification
task. Unlike the sentence-level text classification problem, the available
supervised data for the ERC task is limited, which potentially prevents the
models from playing their maximum effect. In this paper, we propose a novel
approach to leverage unsupervised conversation data, which is more accessible.
Specifically, we propose the Conversation Completion (ConvCom) task, which
attempts to select the correct answer from candidate answers to fill a masked
utterance in a conversation. Then, we Pre-train a basic COntext- Dependent
Encoder (Pre-CODE) on the ConvCom task. Finally, we fine-tune the Pre-CODE on
the datasets of ERC. Experimental results demonstrate that pre-training on
unsupervised data achieves significant improvement of performance on the ERC
datasets, particularly on the minority emotion classes.
Related papers
- RECAP: Retrieval-Enhanced Context-Aware Prefix Encoder for Personalized
Dialogue Response Generation [30.245143345565758]
We propose a new retrieval-enhanced approach for personalized response generation.
We design a hierarchical transformer retriever trained on dialogue domain data to perform personalized retrieval and a context-aware prefix encoder that fuses the retrieved information to the decoder more effectively.
We quantitatively evaluate our model's performance under a suite of human and automatic metrics and find it to be superior compared to state-of-the-art baselines on English Reddit conversations.
arXiv Detail & Related papers (2023-06-12T16:10:21Z) - Socratic Pretraining: Question-Driven Pretraining for Controllable
Summarization [89.04537372465612]
Socratic pretraining is a question-driven, unsupervised pretraining objective designed to improve controllability in summarization tasks.
Our results show that Socratic pretraining cuts task-specific labeled data requirements in half.
arXiv Detail & Related papers (2022-12-20T17:27:10Z) - The Conversational Short-phrase Speaker Diarization (CSSD) Task:
Dataset, Evaluation Metric and Baselines [63.86406909879314]
This paper describes the Conversational Short-phrases Speaker Diarization (CSSD) task.
It consists of training and testing datasets, evaluation metric and baselines.
In the metric aspect, we design the new conversational DER (CDER) evaluation metric, which calculates the SD accuracy at the utterance level.
arXiv Detail & Related papers (2022-08-17T03:26:23Z) - End-to-end Spoken Conversational Question Answering: Task, Dataset and
Model [92.18621726802726]
In spoken question answering, the systems are designed to answer questions from contiguous text spans within the related speech transcripts.
We propose a new Spoken Conversational Question Answering task (SCQA), aiming at enabling the systems to model complex dialogue flows.
Our main objective is to build the system to deal with conversational questions based on the audio recordings, and to explore the plausibility of providing more cues from different modalities with systems in information gathering.
arXiv Detail & Related papers (2022-04-29T17:56:59Z) - Findings on Conversation Disentanglement [28.874162427052905]
We build a learning model that learns utterance-to-utterance and utterance-to-thread classification.
Experiments on the Ubuntu IRC dataset show that this approach has the potential to outperform the conventional greedy approach.
arXiv Detail & Related papers (2021-12-10T05:54:48Z) - Unsupervised Conversation Disentanglement through Co-Training [30.304609312675186]
We explore to train a conversation disentanglement model without referencing any human annotations.
Our method is built upon a deep co-training algorithm, which consists of two neural networks.
For the message-pair classifier, we enrich its training data by retrieving message pairs with high confidence.
arXiv Detail & Related papers (2021-09-07T17:05:18Z) - Unsupervised Summarization for Chat Logs with Topic-Oriented Ranking and
Context-Aware Auto-Encoders [59.038157066874255]
We propose a novel framework called RankAE to perform chat summarization without employing manually labeled data.
RankAE consists of a topic-oriented ranking strategy that selects topic utterances according to centrality and diversity simultaneously.
A denoising auto-encoder is designed to generate succinct but context-informative summaries based on the selected utterances.
arXiv Detail & Related papers (2020-12-14T07:31:17Z) - Towards Data Distillation for End-to-end Spoken Conversational Question
Answering [65.124088336738]
We propose a new Spoken Conversational Question Answering task (SCQA)
SCQA aims at enabling QA systems to model complex dialogues flow given the speech utterances and text corpora.
Our main objective is to build a QA system to deal with conversational questions both in spoken and text forms.
arXiv Detail & Related papers (2020-10-18T05:53:39Z) - Contextual Dialogue Act Classification for Open-Domain Conversational
Agents [10.576497782941697]
Classifying the general intent of the user utterance in a conversation, also known as Dialogue Act (DA), is a key step in Natural Language Understanding (NLU) for conversational agents.
We propose CDAC (Contextual Dialogue Act), a simple yet effective deep learning approach for contextual dialogue act classification.
We use transfer learning to adapt models trained on human-human conversations to predict dialogue acts in human-machine dialogues.
arXiv Detail & Related papers (2020-05-28T06:48:10Z) - ConCET: Entity-Aware Topic Classification for Open-Domain Conversational
Agents [9.870634472479571]
We introduce ConCET: a Concurrent Entity-aware conversational Topic classifier.
We propose a simple and effective method for generating synthetic training data.
We evaluate ConCET on a large dataset of human-machine conversations with real users, collected as part of the Amazon Alexa Prize.
arXiv Detail & Related papers (2020-05-28T06:29:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.