Classification with abstention but without disparities
- URL: http://arxiv.org/abs/2102.12258v1
- Date: Wed, 24 Feb 2021 12:43:55 GMT
- Title: Classification with abstention but without disparities
- Authors: Nicolas Schreuder and Evgenii Chzhen
- Abstract summary: We build a general purpose classification algorithm, which is able to abstain from prediction, while avoiding disparate impact.
We establish finite sample risk, fairness, and abstention guarantees for the proposed algorithm.
Our method empirically shows that moderate abstention rates allow to bypass the risk-fairness trade-off.
- Score: 5.025654873456756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classification with abstention has gained a lot of attention in recent years
as it allows to incorporate human decision-makers in the process. Yet,
abstention can potentially amplify disparities and lead to discriminatory
predictions. The goal of this work is to build a general purpose classification
algorithm, which is able to abstain from prediction, while avoiding disparate
impact. We formalize this problem as risk minimization under fairness and
abstention constraints for which we derive the form of the optimal classifier.
Building on this result, we propose a post-processing classification algorithm,
which is able to modify any off-the-shelf score-based classifier using only
unlabeled sample. We establish finite sample risk, fairness, and abstention
guarantees for the proposed algorithm. In particular, it is shown that fairness
and abstention constraints can be achieved independently from the initial
classifier as long as sufficiently many unlabeled data is available. The risk
guarantee is established in terms of the quality of the initial classifier. Our
post-processing scheme reduces to a sparse linear program allowing for an
efficient implementation, which we provide. Finally, we validate our method
empirically showing that moderate abstention rates allow to bypass the
risk-fairness trade-off.
Related papers
- Fair Classifiers that Abstain without Harm [24.90899074869189]
In critical applications, it is vital for classifiers to defer decision-making to humans.
We propose a post-hoc method that makes existing classifiers selectively abstain from predicting certain samples.
Our framework outperforms existing methods in terms of fairness disparity without sacrificing accuracy at similar abstention rates.
arXiv Detail & Related papers (2023-10-09T23:07:28Z) - Anti-Exploration by Random Network Distillation [63.04360288089277]
We show that a naive choice of conditioning for the Random Network Distillation (RND) is not discriminative enough to be used as an uncertainty estimator.
We show that this limitation can be avoided with conditioning based on Feature-wise Linear Modulation (FiLM)
We evaluate it on the D4RL benchmark, showing that it is capable of achieving performance comparable to ensemble-based methods and outperforming ensemble-free approaches by a wide margin.
arXiv Detail & Related papers (2023-01-31T13:18:33Z) - Neighbour Consistency Guided Pseudo-Label Refinement for Unsupervised
Person Re-Identification [80.98291772215154]
Unsupervised person re-identification (ReID) aims at learning discriminative identity features for person retrieval without any annotations.
Recent advances accomplish this task by leveraging clustering-based pseudo labels.
We propose a Neighbour Consistency guided Pseudo Label Refinement framework.
arXiv Detail & Related papers (2022-11-30T09:39:57Z) - Self-Certifying Classification by Linearized Deep Assignment [65.0100925582087]
We propose a novel class of deep predictors for classifying metric data on graphs within PAC-Bayes risk certification paradigm.
Building on the recent PAC-Bayes literature and data-dependent priors, this approach enables learning posterior distributions on the hypothesis space.
arXiv Detail & Related papers (2022-01-26T19:59:14Z) - Scalable Optimal Classifiers for Adversarial Settings under Uncertainty [10.90668635921398]
We consider the problem of finding optimal classifiers in an adversarial setting where the class-1 data is generated by an attacker whose objective is not known to the defender.
We show that this low-dimensional characterization enables to develop a training method to compute provably approximately optimal classifiers in a scalable manner.
arXiv Detail & Related papers (2021-06-28T13:33:53Z) - Mitigating Bias in Set Selection with Noisy Protected Attributes [16.882719401742175]
We show that in the presence of noisy protected attributes, in attempting to increase fairness without considering noise, one can, in fact, decrease the fairness of the result!
We formulate a denoised'' selection problem which functions for a large class of fairness metrics.
Our empirical results show that this approach can produce subsets which significantly improve the fairness metrics despite the presence of noisy protected attributes.
arXiv Detail & Related papers (2020-11-09T06:45:15Z) - Selective Classification via One-Sided Prediction [54.05407231648068]
One-sided prediction (OSP) based relaxation yields an SC scheme that attains near-optimal coverage in the practically relevant high target accuracy regime.
We theoretically derive bounds generalization for SC and OSP, and empirically we show that our scheme strongly outperforms state of the art methods in coverage at small error levels.
arXiv Detail & Related papers (2020-10-15T16:14:27Z) - Finite-sample Analysis of Interpolating Linear Classifiers in the
Overparameterized Regime [16.1176305285103]
We prove bounds on the population risk of the maximum margin algorithm for two-class linear classification.
We analyze this algorithm applied to random data including misclassification noise.
arXiv Detail & Related papers (2020-04-25T00:06:18Z) - Progressive Identification of True Labels for Partial-Label Learning [112.94467491335611]
Partial-label learning (PLL) is a typical weakly supervised learning problem, where each training instance is equipped with a set of candidate labels among which only one is the true label.
Most existing methods elaborately designed as constrained optimizations that must be solved in specific manners, making their computational complexity a bottleneck for scaling up to big data.
This paper proposes a novel framework of classifier with flexibility on the model and optimization algorithm.
arXiv Detail & Related papers (2020-02-19T08:35:15Z) - Certified Robustness to Label-Flipping Attacks via Randomized Smoothing [105.91827623768724]
Machine learning algorithms are susceptible to data poisoning attacks.
We present a unifying view of randomized smoothing over arbitrary functions.
We propose a new strategy for building classifiers that are pointwise-certifiably robust to general data poisoning attacks.
arXiv Detail & Related papers (2020-02-07T21:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.