On the Fairness Impacts of Private Ensembles Models
- URL: http://arxiv.org/abs/2305.11807v1
- Date: Fri, 19 May 2023 16:43:53 GMT
- Title: On the Fairness Impacts of Private Ensembles Models
- Authors: Cuong Tran, Ferdinando Fioretto
- Abstract summary: Private Aggregation of Teacher Ensembles (PATE) is a machine learning framework that enables the creation of private models.
This paper explores whether the use of PATE can result in unfairness, and demonstrates that it can lead to accuracy disparities among groups of individuals.
- Score: 44.15285550981899
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Private Aggregation of Teacher Ensembles (PATE) is a machine learning
framework that enables the creation of private models through the combination
of multiple "teacher" models and a "student" model. The student model learns to
predict an output based on the voting of the teachers, and the resulting model
satisfies differential privacy. PATE has been shown to be effective in creating
private models in semi-supervised settings or when protecting data labels is a
priority. This paper explores whether the use of PATE can result in unfairness,
and demonstrates that it can lead to accuracy disparities among groups of
individuals. The paper also analyzes the algorithmic and data properties that
contribute to these disproportionate impacts, why these aspects are affecting
different groups disproportionately, and offers recommendations for mitigating
these effects
Related papers
- Is Your Model "MADD"? A Novel Metric to Evaluate Algorithmic Fairness
for Predictive Student Models [0.0]
We propose a novel metric, the Model Absolute Density Distance (MADD), to analyze models' discriminatory behaviors.
We evaluate our approach on the common task of predicting student success in online courses, using several common predictive classification models.
arXiv Detail & Related papers (2023-05-24T16:55:49Z) - Striving for data-model efficiency: Identifying data externalities on
group performance [75.17591306911015]
Building trustworthy, effective, and responsible machine learning systems hinges on understanding how differences in training data and modeling decisions interact to impact predictive performance.
We focus on a particular type of data-model inefficiency, in which adding training data from some sources can actually lower performance evaluated on key sub-groups of the population.
Our results indicate that data-efficiency is a key component of both accurate and trustworthy machine learning.
arXiv Detail & Related papers (2022-11-11T16:48:27Z) - Private Semi-supervised Knowledge Transfer for Deep Learning from Noisy
Labels [21.374707481630697]
We propose PATE++, which combines the current advanced noisy label training mechanisms with the original PATE framework to enhance its accuracy.
A novel structure of Generative Adversarial Nets (GANs) is developed in order to integrate them effectively.
In addition, we develop a novel noisy label detection mechanism for semi-supervised model training to further improve student model performance when training with noisy labels.
arXiv Detail & Related papers (2022-11-03T07:33:49Z) - "You Can't Fix What You Can't Measure": Privately Measuring Demographic
Performance Disparities in Federated Learning [78.70083858195906]
We propose differentially private mechanisms to measure differences in performance across groups while protecting the privacy of group membership.
Our results show that, contrary to what prior work suggested, protecting privacy is not necessarily in conflict with identifying performance disparities of federated models.
arXiv Detail & Related papers (2022-06-24T09:46:43Z) - SF-PATE: Scalable, Fair, and Private Aggregation of Teacher Ensembles [50.90773979394264]
This paper studies a model that protects the privacy of individuals' sensitive information while also allowing it to learn non-discriminatory predictors.
A key characteristic of the proposed model is to enable the adoption of off-the-selves and non-private fair models to create a privacy-preserving and fair model.
arXiv Detail & Related papers (2022-04-11T14:42:54Z) - A Fairness Analysis on Private Aggregation of Teacher Ensembles [31.388212637482365]
The Private Aggregation of Teacher Ensembles (PATE) is an important private machine learning framework.
This paper asks whether this privacy-preserving framework introduces or exacerbates bias and unfairness.
It shows that PATE can introduce accuracy disparity among individuals and groups of individuals.
arXiv Detail & Related papers (2021-09-17T16:19:24Z) - Differentially Private Deep Learning under the Fairness Lens [34.28936739262812]
Differential Privacy (DP) is an important privacy-enhancing technology for private machine learning systems.
It allows to measure and bound the risk associated with an individual participation in a computation.
It was recently observed that DP learning systems may exacerbate bias and unfairness for different groups of individuals.
arXiv Detail & Related papers (2021-06-04T19:10:09Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z) - Differentially Private and Fair Deep Learning: A Lagrangian Dual
Approach [54.32266555843765]
This paper studies a model that protects the privacy of the individuals sensitive information while also allowing it to learn non-discriminatory predictors.
The method relies on the notion of differential privacy and the use of Lagrangian duality to design neural networks that can accommodate fairness constraints.
arXiv Detail & Related papers (2020-09-26T10:50:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.