Participatory Personalization in Classification
- URL: http://arxiv.org/abs/2302.03874v2
- Date: Wed, 11 Oct 2023 23:12:41 GMT
- Title: Participatory Personalization in Classification
- Authors: Hailey Joren, Chirag Nagpal, Katherine Heller, Berk Ustun
- Abstract summary: We introduce a family of classification models, called participatory systems, that let individuals opt into personalization at prediction time.
We conduct a comprehensive empirical study of participatory systems in clinical prediction tasks, benchmarking them with common approaches for personalization and imputation.
Our results demonstrate that participatory systems can facilitate and inform consent while improving performance and data use across all groups who report personal data.
- Score: 8.234011679612436
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning models are often personalized with information that is
protected, sensitive, self-reported, or costly to acquire. These models use
information about people but do not facilitate nor inform their consent.
Individuals cannot opt out of reporting personal information to a model, nor
tell if they benefit from personalization in the first place. We introduce a
family of classification models, called participatory systems, that let
individuals opt into personalization at prediction time. We present a
model-agnostic algorithm to learn participatory systems for personalization
with categorical group attributes. We conduct a comprehensive empirical study
of participatory systems in clinical prediction tasks, benchmarking them with
common approaches for personalization and imputation. Our results demonstrate
that participatory systems can facilitate and inform consent while improving
performance and data use across all groups who report personal data.
Related papers
- PersonaBench: Evaluating AI Models on Understanding Personal Information through Accessing (Synthetic) Private User Data [76.21047984886273]
Personalization is critical in AI assistants, particularly in the context of private AI models that work with individual users.
Due to the sensitive nature of such data, there are no publicly available datasets that allow us to assess an AI model's ability to understand users.
We introduce a synthetic data generation pipeline that creates diverse, realistic user profiles and private documents simulating human activities.
arXiv Detail & Related papers (2025-02-28T00:43:35Z) - When Machine Learning Gets Personal: Understanding Fairness of Personalized Models [5.002195711989324]
Personalization in machine learning involves tailoring models to individual users by incorporating personal attributes such as demographic or medical data.
While personalization can improve prediction accuracy, it may also amplify biases and reduce explainability.
This work introduces a unified framework to evaluate the impact of personalization on both prediction accuracy and explanation quality.
arXiv Detail & Related papers (2025-02-05T00:17:33Z) - Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes [72.13373216644021]
We study the societal impact of machine learning by considering the collection of models that are deployed in a given context.
We find deployed machine learning is prone to systemic failure, meaning some users are exclusively misclassified by all models available.
These examples demonstrate ecosystem-level analysis has unique strengths for characterizing the societal impact of machine learning.
arXiv Detail & Related papers (2023-07-12T01:11:52Z) - IP-FL: Incentivized and Personalized Federated Learning [13.13354915338396]
We first propose to treat incentivization and personalization as interrelated challenges and solve them with an incentive mechanism that fosters personalized learning.
Our approach enhances the personalized model appeal for self-aware clients with high-quality data leading to their active and consistent participation.
arXiv Detail & Related papers (2023-04-15T09:02:06Z) - Visual Prompt Based Personalized Federated Learning [83.04104655903846]
We propose a novel PFL framework for image classification tasks, dubbed pFedPT, that leverages personalized visual prompts to implicitly represent local data distribution information of clients.
Experiments on the CIFAR10 and CIFAR100 datasets show that pFedPT outperforms several state-of-the-art (SOTA) PFL algorithms by a large margin in various settings.
arXiv Detail & Related papers (2023-03-15T15:02:15Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - "You Can't Fix What You Can't Measure": Privately Measuring Demographic
Performance Disparities in Federated Learning [78.70083858195906]
We propose differentially private mechanisms to measure differences in performance across groups while protecting the privacy of group membership.
Our results show that, contrary to what prior work suggested, protecting privacy is not necessarily in conflict with identifying performance disparities of federated models.
arXiv Detail & Related papers (2022-06-24T09:46:43Z) - SF-PATE: Scalable, Fair, and Private Aggregation of Teacher Ensembles [50.90773979394264]
This paper studies a model that protects the privacy of individuals' sensitive information while also allowing it to learn non-discriminatory predictors.
A key characteristic of the proposed model is to enable the adoption of off-the-selves and non-private fair models to create a privacy-preserving and fair model.
arXiv Detail & Related papers (2022-04-11T14:42:54Z) - UserIdentifier: Implicit User Representations for Simple and Effective
Personalized Sentiment Analysis [36.162520010250056]
We propose UserIdentifier, a novel scheme for training a single shared model for all users.
Our approach produces personalized responses by adding fixed, non-trainable user identifiers to the input data.
arXiv Detail & Related papers (2021-10-01T00:21:33Z) - Opportunistic Federated Learning: An Exploration of Egocentric
Collaboration for Pervasive Computing Applications [20.61034787249924]
We define a new approach, opportunistic federated learning, in which individual devices belonging to different users seek to learn robust models.
In this paper, we explore the feasibility and limits of such an approach, culminating in a framework that supports encounter-based pairwise collaborative learning.
arXiv Detail & Related papers (2021-03-24T15:30:21Z) - Fairness-Aware Online Personalization [16.320648868892526]
We present a study of fairness in online personalization settings involving the ranking of individuals.
We first demonstrate that online personalization can cause the model to learn to act in an unfair manner if the user is biased in his/her responses.
We then formulate the problem of learning personalized models under fairness constraints and present a regularization based approach for mitigating biases in machine learning.
arXiv Detail & Related papers (2020-07-30T07:16:17Z) - Survey of Personalization Techniques for Federated Learning [0.08594140167290096]
Federated learning enables machine learning models to learn from private decentralized data without compromising privacy.
This paper highlights the need for personalization and surveys recent research on this topic.
arXiv Detail & Related papers (2020-03-19T10:47:55Z) - Three Approaches for Personalization with Applications to Federated
Learning [68.19709953755238]
We present a systematic learning-theoretic study of personalization.
We provide learning-theoretic guarantees and efficient algorithms for which we also demonstrate the performance.
All of our algorithms are model-agnostic and work for any hypothesis class.
arXiv Detail & Related papers (2020-02-25T01:36:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.