An example of prediction which complies with Demographic Parity and
equalizes group-wise risks in the context of regression
- URL: http://arxiv.org/abs/2011.07158v1
- Date: Fri, 13 Nov 2020 22:46:05 GMT
- Title: An example of prediction which complies with Demographic Parity and
equalizes group-wise risks in the context of regression
- Authors: Evgenii Chzhen and Nicolas Schreuder
- Abstract summary: Bayes optimal prediction $f*$ which does not produce Disparate Treatment is defined as $f*(x) = mathbbE[Y | X = x]$.
We discuss several implications of this result on better understanding of mathematical notions of algorithmic fairness.
- Score: 3.9596068699962323
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Let $(X, S, Y) \in \mathbb{R}^p \times \{1, 2\} \times \mathbb{R}$ be a
triplet following some joint distribution $\mathbb{P}$ with feature vector $X$,
sensitive attribute $S$ , and target variable $Y$. The Bayes optimal prediction
$f^*$ which does not produce Disparate Treatment is defined as $f^*(x) =
\mathbb{E}[Y | X = x]$. We provide a non-trivial example of a prediction $x \to
f(x)$ which satisfies two common group-fairness notions: Demographic Parity
\begin{align} (f(X) | S = 1) &\stackrel{d}{=} (f(X) | S = 2) \end{align} and
Equal Group-Wise Risks \begin{align}
\mathbb{E}[(f^*(X) - f(X))^2 | S = 1] = \mathbb{E}[(f^*(X) - f(X))^2 | S =
2]. \end{align} To the best of our knowledge this is the first explicit
construction of a non-constant predictor satisfying the above. We discuss
several implications of this result on better understanding of mathematical
notions of algorithmic fairness.
Related papers
- Learning a Single Neuron Robustly to Distributional Shifts and Adversarial Label Noise [38.551072383777594]
We study the problem of learning a single neuron with respect to the $L2$ loss in the presence of adversarial distribution shifts.
A new algorithm is developed to approximate the vector vector squared loss with respect to the worst distribution that is in the $chi2$divergence to the $mathcalp_0$.
arXiv Detail & Related papers (2024-11-11T03:43:52Z) - The Communication Complexity of Approximating Matrix Rank [50.6867896228563]
We show that this problem has randomized communication complexity $Omega(frac1kcdot n2log|mathbbF|)$.
As an application, we obtain an $Omega(frac1kcdot n2log|mathbbF|)$ space lower bound for any streaming algorithm with $k$ passes.
arXiv Detail & Related papers (2024-10-26T06:21:42Z) - In-depth Analysis of Low-rank Matrix Factorisation in a Federated Setting [21.002519159190538]
We analyze a distributed algorithm to compute a low-rank matrix factorization on $N$ clients.
We obtain a global $mathbfV$ in $mathbbRd times r$ common to all clients and a local $mathbfUi$ in $mathbbRn_itimes r$.
arXiv Detail & Related papers (2024-09-13T12:28:42Z) - Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Conformal prediction under ambiguous ground truth [34.92407089333591]
Conformal Prediction (CP) allows to perform rigorous uncertainty quantification.
It is typically assumed that $mathbbPY|X$ is the "true" posterior label distribution.
We develop Monte Carlo CP procedures which provide guarantees w.r.t. $mathbbP_vote=mathbbPX otimes mathbbP_voteY|X$.
arXiv Detail & Related papers (2023-07-18T14:40:48Z) - Learning a Single Neuron with Adversarial Label Noise via Gradient
Descent [50.659479930171585]
We study a function of the form $mathbfxmapstosigma(mathbfwcdotmathbfx)$ for monotone activations.
The goal of the learner is to output a hypothesis vector $mathbfw$ that $F(mathbbw)=C, epsilon$ with high probability.
arXiv Detail & Related papers (2022-06-17T17:55:43Z) - Near-Optimal Model Discrimination with Non-Disclosure [19.88145627448243]
We first consider the case of a well-specified linear model with squared loss.
We derive sample complexity of a similar form, even under misspecification.
arXiv Detail & Related papers (2020-12-04T23:52:54Z) - Optimal Mean Estimation without a Variance [103.26777953032537]
We study the problem of heavy-tailed mean estimation in settings where the variance of the data-generating distribution does not exist.
We design an estimator which attains the smallest possible confidence interval as a function of $n,d,delta$.
arXiv Detail & Related papers (2020-11-24T22:39:21Z) - Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and
ReLUs under Gaussian Marginals [49.60752558064027]
We study the fundamental problems of agnostically learning halfspaces and ReLUs under Gaussian marginals.
Our lower bounds provide strong evidence that current upper bounds for these tasks are essentially best possible.
arXiv Detail & Related papers (2020-06-29T17:10:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.