Generalization and Personalization of Mobile Sensing-Based Mood
Inference Models: An Analysis of College Students in Eight Countries
- URL: http://arxiv.org/abs/2211.03009v1
- Date: Sun, 6 Nov 2022 02:26:52 GMT
- Title: Generalization and Personalization of Mobile Sensing-Based Mood
Inference Models: An Analysis of College Students in Eight Countries
- Authors: Lakmal Meegahapola, William Droz, Peter Kun, Amalia de Gotzen,
Chaitanya Nutakki, Shyam Diwakar, Salvador Ruiz Correa, Donglei Song, Hao Xu,
Miriam Bidoglia, George Gaskell, Altangerel Chagnaa, Amarsanaa Ganbold,
Tsolmon Zundui, Carlo Caprini, Daniele Miorandi, Alethia Hume, Jose Luis
Zarza, Luca Cernuzzi, Ivano Bison, Marcelo Rodas Britez, Matteo Busso, Ronald
Chenu-Abente, Can Gunel, Fausto Giunchiglia, Laura Schelenz, and Daniel
Gatica-Perez
- Abstract summary: We collect a mobile sensing dataset with 329K self-reports from 678 participants in eight countries.
We evaluate country-specific (trained and tested within a country), continent-specific (trained and tested within a continent), country-agnostic (tested on a country not seen on training data) and multi-country (trained and tested with multiple countries) approaches.
- Score: 8.218081835111912
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mood inference with mobile sensing data has been studied in ubicomp
literature over the last decade. This inference enables context-aware and
personalized user experiences in general mobile apps and valuable feedback and
interventions in mobile health apps. However, even though model generalization
issues have been highlighted in many studies, the focus has always been on
improving the accuracies of models using different sensing modalities and
machine learning techniques, with datasets collected in homogeneous
populations. In contrast, less attention has been given to studying the
performance of mood inference models to assess whether models generalize to new
countries. In this study, we collected a mobile sensing dataset with 329K
self-reports from 678 participants in eight countries (China, Denmark, India,
Italy, Mexico, Mongolia, Paraguay, UK) to assess the effect of geographical
diversity on mood inference models. We define and evaluate country-specific
(trained and tested within a country), continent-specific (trained and tested
within a continent), country-agnostic (tested on a country not seen on training
data), and multi-country (trained and tested with multiple countries)
approaches trained on sensor data for two mood inference tasks with
population-level (non-personalized) and hybrid (partially personalized) models.
We show that partially personalized country-specific models perform the best
yielding area under the receiver operating characteristic curve (AUROC) scores
of the range 0.78-0.98 for two-class (negative vs. positive valence) and
0.76-0.94 for three-class (negative vs. neutral vs. positive valence)
inference. Overall, we uncover generalization issues of mood inference models
to new countries and how the geographical similarity of countries might impact
mood inference.
Related papers
- Towards Geographic Inclusion in the Evaluation of Text-to-Image Models [25.780536950323683]
We study how much annotators in Africa, Europe, and Southeast Asia vary in their perception of geographic representation, visual appeal, and consistency in real and generated images.
For example, annotators in different locations often disagree on whether exaggerated, stereotypical depictions of a region are considered geographically representative.
We recommend steps for improved automatic and human evaluations.
arXiv Detail & Related papers (2024-05-07T16:23:06Z) - The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models [67.38144169029617]
We introduce PRISM, a new dataset which maps the sociodemographics and stated preferences of 1,500 diverse participants from 75 countries.
PRISM contributes (i) wide geographic and demographic participation in human feedback data; (ii) two census-representative samples for understanding collective welfare (UK and US); and (iii) individualised feedback where every rating is linked to a detailed participant profile.
arXiv Detail & Related papers (2024-04-24T17:51:36Z) - Exposing Bias in Online Communities through Large-Scale Language Models [3.04585143845864]
This work uses the flaw of bias in language models to explore the biases of six different online communities.
The bias of the resulting models is evaluated by prompting the models with different demographics and comparing the sentiment and toxicity values of these generations.
This work not only affirms how easily bias is absorbed from training data but also presents a scalable method to identify and compare the bias of different datasets or communities.
arXiv Detail & Related papers (2023-06-04T08:09:26Z) - Learning About Social Context from Smartphone Data: Generalization
Across Countries and Daily Life Moments [5.764112063319108]
We used a novel, large-scale, and multimodal smartphone sensing dataset with over 216K self-reports collected from 581 young adults in five countries.
Several sensors are informative of social context, that partially personalized multi-country models (trained and tested with data from all countries) and country-specific models (trained and tested within countries) can achieve similar performance above 90% AUC.
These findings confirm the importance of the diversity of mobile data, to better understand social context inference models in different countries.
arXiv Detail & Related papers (2023-06-01T17:20:56Z) - Inferring Mood-While-Eating with Smartphone Sensing and Community-Based
Model Personalization [4.245223529153532]
Phone sensor data have been used to characterize both eating behavior and mood, independently, in the context of mobile food diaries and mobile health applications.
Our results indicate that generic mood inference models decline in performance in certain contexts, such as when eating.
To address these limitations, we employed a novel community-based approach for personalization by building models with data from a set of similar users to a target user.
arXiv Detail & Related papers (2023-06-01T14:24:10Z) - Bridging the Gap: A Survey on Integrating (Human) Feedback for Natural
Language Generation [68.9440575276396]
This survey aims to provide an overview of the recent research that has leveraged human feedback to improve natural language generation.
First, we introduce an encompassing formalization of feedback, and identify and organize existing research into a taxonomy following this formalization.
Second, we discuss how feedback can be described by its format and objective, and cover the two approaches proposed to use feedback (either for training or decoding): directly using the feedback or training feedback models.
Third, we provide an overview of the nascent field of AI feedback, which exploits large language models to make judgments based on a set of principles and minimize the need for
arXiv Detail & Related papers (2023-05-01T17:36:06Z) - Fairness meets Cross-Domain Learning: a new perspective on Models and
Metrics [80.07271410743806]
We study the relationship between cross-domain learning (CD) and model fairness.
We introduce a benchmark on face and medical images spanning several demographic groups as well as classification and localization tasks.
Our study covers 14 CD approaches alongside three state-of-the-art fairness algorithms and shows how the former can outperform the latter.
arXiv Detail & Related papers (2023-03-25T09:34:05Z) - On the limits of algorithmic prediction across the globe [4.392517231156947]
We show that state-of-the-art machine learning models trained on data from the United States can predict achievement with high accuracy and generalize to other developed countries with comparable accuracy.
Training the same model on national data yields high accuracy in every country, which highlights the value of local data collection.
arXiv Detail & Related papers (2021-03-28T19:53:18Z) - LOGAN: Local Group Bias Detection by Clustering [86.38331353310114]
We argue that evaluating bias at the corpus level is not enough for understanding how biases are embedded in a model.
We propose LOGAN, a new bias detection technique based on clustering.
Experiments on toxicity classification and object classification tasks show that LOGAN identifies bias in a local region.
arXiv Detail & Related papers (2020-10-06T16:42:51Z) - Dataset Cartography: Mapping and Diagnosing Datasets with Training
Dynamics [118.75207687144817]
We introduce Data Maps, a model-based tool to characterize and diagnose datasets.
We leverage a largely ignored source of information: the behavior of the model on individual instances during training.
Our results indicate that a shift in focus from quantity to quality of data could lead to robust models and improved out-of-distribution generalization.
arXiv Detail & Related papers (2020-09-22T20:19:41Z) - Think Locally, Act Globally: Federated Learning with Local and Global
Representations [92.68484710504666]
Federated learning is a method of training models on private data distributed over multiple devices.
We propose a new federated learning algorithm that jointly learns compact local representations on each device.
We also evaluate on the task of personalized mood prediction from real-world mobile data where privacy is key.
arXiv Detail & Related papers (2020-01-06T12:40:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.