Predicting User Code-Switching Level from Sociological and Psychological
Profiles
- URL: http://arxiv.org/abs/2112.06462v1
- Date: Mon, 13 Dec 2021 07:36:02 GMT
- Title: Predicting User Code-Switching Level from Sociological and Psychological
Profiles
- Authors: Injy Hamed, Alia El Bolock, Nader Rizk, Cornelia Herbert, Slim
Abdennadher, Ngoc Thang Vu
- Abstract summary: We show the correlation between users' CS frequency and character traits.
We use machine learning (ML) to validate the findings.
The predictive models were able to predict users' CS frequency with an accuracy higher than 55%.
- Score: 24.32063659777203
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multilingual speakers tend to alternate between languages within a
conversation, a phenomenon referred to as "code-switching" (CS). CS is a
complex phenomenon that not only encompasses linguistic challenges, but also
contains a great deal of complexity in terms of its dynamic behaviour across
speakers. This dynamic behaviour has been studied by sociologists and
psychologists, identifying factors affecting CS. In this paper, we provide an
empirical user study on Arabic-English CS, where we show the correlation
between users' CS frequency and character traits. We use machine learning (ML)
to validate the findings, informing and confirming existing theories. The
predictive models were able to predict users' CS frequency with an accuracy
higher than 55%, where travel experiences and personality traits played the
biggest role in the modeling process.
Related papers
- On the Proper Treatment of Tokenization in Psycholinguistics [53.960910019072436]
The paper argues that token-level language models should be marginalized into character-level language models before they are used in psycholinguistic studies.
We find various focal areas whose surprisal is a better psychometric predictor than the surprisal of the region of interest itself.
arXiv Detail & Related papers (2024-10-03T17:18:03Z) - Unveiling Language Competence Neurons: A Psycholinguistic Approach to Model Interpretability [2.672177830116334]
This study employs psycholinguistic paradigms to explore neuron-level representations in language model across three tasks.
Our findings indicate that while GPT-2-XL struggles with the sound-shape task, it demonstrates human-like abilities in both sound-gender association and implicit causality.
arXiv Detail & Related papers (2024-09-24T07:40:33Z) - PersLLM: A Personified Training Approach for Large Language Models [66.16513246245401]
We propose PersLLM, integrating psychology-grounded principles of personality: social practice, consistency, and dynamic development.
We incorporate personality traits directly into the model parameters, enhancing the model's resistance to induction, promoting consistency, and supporting the dynamic evolution of personality.
arXiv Detail & Related papers (2024-07-17T08:13:22Z) - Beyond Text: Leveraging Multi-Task Learning and Cognitive Appraisal Theory for Post-Purchase Intention Analysis [10.014248704653]
This study evaluates multi-task learning frameworks grounded in Cognitive Appraisal Theory to predict user behavior.
Our experiments show that users' language and traits improve predictions above and beyond models predicting only from text.
arXiv Detail & Related papers (2024-07-11T04:57:52Z) - Modulating Language Model Experiences through Frictions [56.17593192325438]
Over-consumption of language model outputs risks propagating unchecked errors in the short-term and damaging human capabilities for critical thinking in the long-term.
We propose selective frictions for language model experiences, inspired by behavioral science interventions, to dampen misuse.
arXiv Detail & Related papers (2024-06-24T16:31:11Z) - The Who in Code-Switching: A Case Study for Predicting Egyptian
Arabic-English Code-Switching Levels based on Character Profiles [20.746558640332953]
Code-switching (CS) is a linguistic phenomenon exhibited by multilingual individuals, where they tend to alternate between languages within one single conversation.
We use machine learning (ML) to predict users' CS levels based on their profiles.
Our results show that the CS behaviour is affected by the relation between speakers, travel experiences as well as Neuroticism and Extraversion personality traits.
arXiv Detail & Related papers (2022-07-31T13:47:35Z) - Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A
Survey [71.43956423427397]
We aim to identify the nonverbal cues and computational methodologies resulting in effective performance.
This survey differs from its counterparts by involving the widest spectrum of social phenomena and interaction settings.
Some major observations are: the most often used nonverbal cue, computational method, interaction environment, and sensing approach are speaking activity, support vector machines, and meetings composed of 3-4 persons equipped with microphones and cameras, respectively.
arXiv Detail & Related papers (2022-07-20T13:37:57Z) - Perception Point: Identifying Critical Learning Periods in Speech for
Bilingual Networks [58.24134321728942]
We compare and identify cognitive aspects on deep neural-based visual lip-reading models.
We observe a strong correlation between these theories in cognitive psychology and our unique modeling.
arXiv Detail & Related papers (2021-10-13T05:30:50Z) - BERT meets LIWC: Exploring State-of-the-Art Language Models for
Predicting Communication Behavior in Couples' Conflict Interactions [3.0309575462589122]
We train machine learning models to automatically predict communication codes of 368 German-speaking Swiss couples.
Results suggest it might be time to consider modern alternatives to LIWC, the de facto linguistic features in psychology.
arXiv Detail & Related papers (2021-06-03T01:37:59Z) - Style Variation as a Vantage Point for Code-Switching [54.34370423151014]
Code-Switching (CS) is a common phenomenon observed in several bilingual and multilingual communities.
We present a novel vantage point of CS to be style variations between both the participating languages.
We propose a two-stage generative adversarial training approach where the first stage generates competitive negative examples for CS and the second stage generates more realistic CS sentences.
arXiv Detail & Related papers (2020-05-01T15:53:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.