BERT-based Ensembles for Modeling Disclosure and Support in
Conversational Social Media Text
- URL: http://arxiv.org/abs/2006.01222v1
- Date: Mon, 1 Jun 2020 19:52:01 GMT
- Title: BERT-based Ensembles for Modeling Disclosure and Support in
Conversational Social Media Text
- Authors: Tanvi Dadu, Kartikey Pant and Radhika Mamidi
- Abstract summary: We introduce a predictive ensemble model exploiting the finetuned contextualized word embeddings, RoBERTa and ALBERT.
We show that our model outperforms the base models in all considered metrics, achieving an improvement of $3%$ in the F1 score.
- Score: 9.475039534437332
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There is a growing interest in understanding how humans initiate and hold
conversations. The affective understanding of conversations focuses on the
problem of how speakers use emotions to react to a situation and to each other.
In the CL-Aff Shared Task, the organizers released Get it #OffMyChest dataset,
which contains Reddit comments from casual and confessional conversations,
labeled for their disclosure and supportiveness characteristics. In this paper,
we introduce a predictive ensemble model exploiting the finetuned
contextualized word embeddings, RoBERTa and ALBERT. We show that our model
outperforms the base models in all considered metrics, achieving an improvement
of $3\%$ in the F1 score. We further conduct statistical analysis and outline
deeper insights into the given dataset while providing a new characterization
of impact for the dataset.
Related papers
- Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation [65.16137964758612]
We explore the use of long-context capabilities in large language models to create synthetic reading comprehension data from entire books.
Our objective is to test the capabilities of LLMs to analyze, understand, and reason over problems that require a detailed comprehension of long spans of text.
arXiv Detail & Related papers (2024-05-31T20:15:10Z) - Learning From Free-Text Human Feedback -- Collect New Datasets Or Extend
Existing Ones? [57.16050211534735]
We investigate the types and frequency of free-text human feedback in commonly used dialog datasets.
Our findings provide new insights into the composition of the datasets examined, including error types, user response types, and the relations between them.
arXiv Detail & Related papers (2023-10-24T12:01:11Z) - Rating Prediction in Conversational Task Assistants with Behavioral and
Conversational-Flow Features [6.188306785668896]
We propose TB-Rater, a Transformer model which combines conversational-flow features with user behavior features for predicting user ratings.
In particular, we use real human-agent conversations and ratings collected in the Alexa TaskBot challenge.
Our results show the advantages of modeling both the conversational-flow and behavioral aspects of the conversation in a single model for offline rating prediction.
arXiv Detail & Related papers (2023-09-20T13:34:03Z) - Measuring the Effect of Influential Messages on Varying Personas [67.1149173905004]
We present a new task, Response Forecasting on Personas for News Media, to estimate the response a persona might have upon seeing a news message.
The proposed task not only introduces personalization in the modeling but also predicts the sentiment polarity and intensity of each response.
This enables more accurate and comprehensive inference on the mental state of the persona.
arXiv Detail & Related papers (2023-05-25T21:01:00Z) - TANet: Thread-Aware Pretraining for Abstractive Conversational
Summarization [27.185068253347257]
We build a large-scale (11M) pretraining dataset called RCS based on the multi-person discussions in the Reddit community.
We then present TANet, a thread-aware Transformer-based network.
Unlike the existing pre-trained models that treat a conversation as a sequence of sentences, we argue that the inherent contextual dependency plays an essential role in understanding the entire conversation.
arXiv Detail & Related papers (2022-04-09T16:08:46Z) - T-BERT -- Model for Sentiment Analysis of Micro-blogs Integrating Topic
Model and BERT [0.0]
The effectiveness of BERT(Bidirectional Representations from Transformers) in sentiment classification tasks from a raw live dataset is demonstrated.
A novel T-BERT framework is proposed to show the enhanced performance obtainable by combining latent topics with contextual BERT embeddings.
arXiv Detail & Related papers (2021-06-02T12:01:47Z) - ConvoSumm: Conversation Summarization Benchmark and Improved Abstractive
Summarization with Argument Mining [61.82562838486632]
We crowdsource four new datasets on diverse online conversation forms of news comments, discussion forums, community question answering forums, and email threads.
We benchmark state-of-the-art models on our datasets and analyze characteristics associated with the data.
arXiv Detail & Related papers (2021-06-01T22:17:13Z) - Exploiting Unsupervised Data for Emotion Recognition in Conversations [76.01690906995286]
Emotion Recognition in Conversations (ERC) aims to predict the emotional state of speakers in conversations.
The available supervised data for the ERC task is limited.
We propose a novel approach to leverage unsupervised conversation data.
arXiv Detail & Related papers (2020-10-02T13:28:47Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - DialBERT: A Hierarchical Pre-Trained Model for Conversation
Disentanglement [47.403802900555576]
We propose a new model, named Dialogue BERT (DialBERT), which integrates local and global semantics in a single stream of messages to disentangle the conversations that mixed together.
We employ BERT to capture the matching information in each utterance pair at the utterance-level, and use a BiLSTM to aggregate and incorporate the context-level information.
With only a 3% increase in parameters, a 12% improvement has been attained in comparison to BERT, based on the F1-Score.
arXiv Detail & Related papers (2020-04-08T00:54:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.