Pushing on Personality Detection from Verbal Behavior: A Transformer
Meets Text Contours of Psycholinguistic Features
- URL: http://arxiv.org/abs/2204.04629v1
- Date: Sun, 10 Apr 2022 08:08:46 GMT
- Title: Pushing on Personality Detection from Verbal Behavior: A Transformer
Meets Text Contours of Psycholinguistic Features
- Authors: Elma Kerz, Yu Qiao, Sourabh Zanwar, Daniel Wiechmann
- Abstract summary: We report two major improvements in predicting personality traits from text data.
We integrate a pre-trained Transformer Language Model BERT and Bidirectional Long Short-Term Memory networks trained on within-text distributions of psycholinguistic features.
We evaluate the performance of the models we built on two benchmark datasets.
- Score: 27.799032561722893
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Research at the intersection of personality psychology, computer science, and
linguistics has recently focused increasingly on modeling and predicting
personality from language use. We report two major improvements in predicting
personality traits from text data: (1) to our knowledge, the most comprehensive
set of theory-based psycholinguistic features and (2) hybrid models that
integrate a pre-trained Transformer Language Model BERT and Bidirectional Long
Short-Term Memory (BLSTM) networks trained on within-text distributions ('text
contours') of psycholinguistic features. We experiment with BLSTM models (with
and without Attention) and with two techniques for applying pre-trained
language representations from the transformer model - 'feature-based' and
'fine-tuning'. We evaluate the performance of the models we built on two
benchmark datasets that target the two dominant theoretical models of
personality: the Big Five Essay dataset and the MBTI Kaggle dataset. Our
results are encouraging as our models outperform existing work on the same
datasets. More specifically, our models achieve improvement in classification
accuracy by 2.9% on the Essay dataset and 8.28% on the Kaggle MBTI dataset. In
addition, we perform ablation experiments to quantify the impact of different
categories of psycholinguistic features in the respective personality
prediction models.
Related papers
- Machine Mindset: An MBTI Exploration of Large Language Models [28.2342069623478]
We present a novel approach for integrating Myers-Briggs Type Indicator (MBTI) personality traits into large language models (LLMs)
Our method, "Machine Mindset," involves a two-phase fine-tuning and Direct Preference Optimization (DPO) to embed MBTI traits into LLMs.
arXiv Detail & Related papers (2023-12-20T12:59:31Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - Extensive Evaluation of Transformer-based Architectures for Adverse Drug
Events Extraction [6.78974856327994]
Adverse Event (ADE) extraction is one of the core tasks in digital pharmacovigilance.
We evaluate 19 Transformer-based models for ADE extraction on informal texts.
At the end of our analyses, we identify a list of take-home messages that can be derived from the experimental data.
arXiv Detail & Related papers (2023-06-08T15:25:24Z) - Exploring the Efficacy of Pre-trained Checkpoints in Text-to-Music
Generation Task [86.72661027591394]
We generate complete and semantically consistent symbolic music scores from text descriptions.
We explore the efficacy of using publicly available checkpoints for natural language processing in the task of text-to-music generation.
Our experimental results show that the improvement from using pre-trained checkpoints is statistically significant in terms of BLEU score and edit distance similarity.
arXiv Detail & Related papers (2022-11-21T07:19:17Z) - Myers-Briggs personality classification from social media text using
pre-trained language models [0.0]
We describe a series of experiments in which the well-known Bidirectional Representations from Transformers (BERT) model is fine-tuned to perform MBTI classification.
Our main findings suggest that the current approach significantly outperforms well-known text classification models based on bag-of-words and static word embeddings alike.
arXiv Detail & Related papers (2022-07-10T14:38:09Z) - A Generative Language Model for Few-shot Aspect-Based Sentiment Analysis [90.24921443175514]
We focus on aspect-based sentiment analysis, which involves extracting aspect term, category, and predicting their corresponding polarities.
We propose to reformulate the extraction and prediction tasks into the sequence generation task, using a generative language model with unidirectional attention.
Our approach outperforms the previous state-of-the-art (based on BERT) on average performance by a large margins in few-shot and full-shot settings.
arXiv Detail & Related papers (2022-04-11T18:31:53Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Comparing Test Sets with Item Response Theory [53.755064720563]
We evaluate 29 datasets using predictions from 18 pretrained Transformer models on individual test examples.
We find that Quoref, HellaSwag, and MC-TACO are best suited for distinguishing among state-of-the-art models.
We also observe span selection task format, which is used for QA datasets like QAMR or SQuAD2.0, is effective in differentiating between strong and weak models.
arXiv Detail & Related papers (2021-06-01T22:33:53Z) - Extending the Abstraction of Personality Types based on MBTI with
Machine Learning and Natural Language Processing [0.0]
A data-centric approach with Natural Language Processing (NLP) to predict personality types based on the MBTI.
The experimentation had a robust baseline of stacked models.
The results showed that attention to the data iteration loop focused on quality, explanatory power and representativeness for the abstraction of more relevant/important resources.
arXiv Detail & Related papers (2021-05-25T10:00:16Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Personality Trait Detection Using Bagged SVM over BERT Word Embedding
Ensembles [10.425280599592865]
We present a novel deep learning-based approach for automated personality detection from text.
We leverage state of the art advances in natural language understanding, namely the BERT language model to extract contextualized word embeddings.
Our model outperforms the previous state of the art by 1.04% and, at the same time is significantly more computationally efficient to train.
arXiv Detail & Related papers (2020-10-03T09:25:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.