Sample Selection for Fair and Robust Training
- URL: http://arxiv.org/abs/2110.14222v1
- Date: Wed, 27 Oct 2021 07:17:29 GMT
- Title: Sample Selection for Fair and Robust Training
- Authors: Yuji Roh, Kangwook Lee, Steven Euijong Whang, Changho Suh
- Abstract summary: We propose a sample selection-based algorithm for fair and robust training.
We show that our algorithm obtains fairness and robustness better than or comparable to the state-of-the-art technique.
- Score: 28.94276265328868
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fairness and robustness are critical elements of Trustworthy AI that need to
be addressed together. Fairness is about learning an unbiased model while
robustness is about learning from corrupted data, and it is known that
addressing only one of them may have an adverse affect on the other. In this
work, we propose a sample selection-based algorithm for fair and robust
training. To this end, we formulate a combinatorial optimization problem for
the unbiased selection of samples in the presence of data corruption. Observing
that solving this optimization problem is strongly NP-hard, we propose a greedy
algorithm that is efficient and effective in practice. Experiments show that
our algorithm obtains fairness and robustness that are better than or
comparable to the state-of-the-art technique, both on synthetic and benchmark
real datasets. Moreover, unlike other fair and robust training baselines, our
algorithm can be used by only modifying the sampling step in batch selection
without changing the training algorithm or leveraging additional clean data.
Related papers
- Fast & Fair: Efficient Second-Order Robust Optimization for Fairness in
Machine Learning [0.0]
This project explores adversarial training techniques to develop fairer Deep Neural Networks (DNNs)
DNNs are susceptible to inheriting bias with respect to sensitive attributes such as race and gender, which can lead to life-altering outcomes.
We propose a robust optimization problem, which we demonstrate can improve fairness in several datasets.
arXiv Detail & Related papers (2024-01-04T01:02:55Z) - Bandit-Driven Batch Selection for Robust Learning under Label Noise [20.202806541218944]
We introduce a novel approach for batch selection in Gradient Descent (SGD) training, leveraging bandit algorithms.
Our methodology focuses on optimizing the learning process in the presence of label noise, a prevalent issue in real-world datasets.
arXiv Detail & Related papers (2023-10-31T19:19:01Z) - Towards Robust Dataset Learning [90.2590325441068]
We propose a principled, tri-level optimization to formulate the robust dataset learning problem.
Under an abstraction model that characterizes robust vs. non-robust features, the proposed method provably learns a robust dataset.
arXiv Detail & Related papers (2022-11-19T17:06:10Z) - Robust Fairness-aware Learning Under Sample Selection Bias [17.09665420515772]
We propose a framework for robust and fair learning under sample selection bias.
We develop two algorithms to handle sample selection bias when test data is both available and unavailable.
arXiv Detail & Related papers (2021-05-24T23:23:36Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Low-Regret Active learning [64.36270166907788]
We develop an online learning algorithm for identifying unlabeled data points that are most informative for training.
At the core of our work is an efficient algorithm for sleeping experts that is tailored to achieve low regret on predictable (easy) instances.
arXiv Detail & Related papers (2021-04-06T22:53:45Z) - Metrics and methods for a systematic comparison of fairness-aware
machine learning algorithms [0.0]
This study is the most comprehensive of its kind.
It considers fairness, predictive-performance, calibration quality, and speed of 28 different modelling pipelines.
We also found that fairness-aware algorithms can induce fairness without material drops in predictive power.
arXiv Detail & Related papers (2020-10-08T13:58:09Z) - Beyond Individual and Group Fairness [90.4666341812857]
We present a new data-driven model of fairness that is guided by the unfairness complaints received by the system.
Our model supports multiple fairness criteria and takes into account their potential incompatibilities.
arXiv Detail & Related papers (2020-08-21T14:14:44Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.