Learning by Grouping: A Multilevel Optimization Framework for Improving
Fairness in Classification without Losing Accuracy
- URL: http://arxiv.org/abs/2304.00486v1
- Date: Sun, 2 Apr 2023 08:45:08 GMT
- Title: Learning by Grouping: A Multilevel Optimization Framework for Improving
Fairness in Classification without Losing Accuracy
- Authors: Ramtin Hosseini, Li Zhang, Bhanu Garg, Pengtao Xie
- Abstract summary: In some cases, AI systems can be unfair by exhibiting bias or discrimination against certain social groups.
We propose a novel machine learning framework where the ML model learns to group a diverse set of problems into distinct subgroups to solve each subgroup.
Our proposed framework involves three stages of learning, which are formulated as a three-level optimization problem.
- Score: 19.84719054826755
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The integration of machine learning models in various real-world applications
is becoming more prevalent to assist humans in their daily decision-making
tasks as a result of recent advancements in this field. However, it has been
discovered that there is a tradeoff between the accuracy and fairness of these
decision-making tasks. In some cases, these AI systems can be unfair by
exhibiting bias or discrimination against certain social groups, which can have
severe consequences in real life. Inspired by one of the most well-known human
learning skills called grouping, we address this issue by proposing a novel
machine learning framework where the ML model learns to group a diverse set of
problems into distinct subgroups to solve each subgroup using its specific
sub-model. Our proposed framework involves three stages of learning, which are
formulated as a three-level optimization problem: (i) learning to group
problems into different subgroups; (ii) learning group-specific sub-models for
problem-solving; and (iii) updating group assignments of training examples by
minimizing the validation loss. These three learning stages are performed
end-to-end in a joint manner using gradient descent. To improve fairness and
accuracy, we develop an efficient optimization algorithm to solve this
three-level optimization problem. To further reduce the risk of overfitting in
small datasets, we incorporate domain adaptation techniques in the second stage
of training. We further apply our method to neural architecture search.
Extensive experiments on various datasets demonstrate our method's
effectiveness and performance improvements in both fairness and accuracy. Our
proposed Learning by Grouping can reduce overfitting and achieve
state-of-the-art performances with fixed human-designed network architectures
and searchable network architectures on various datasets.
Related papers
- A study on a Q-Learning algorithm application to a manufacturing
assembly problem [0.8937905773981699]
This study focuses on the implementation of a reinforcement learning algorithm in an assembly problem of a given object.
A model-free Q-Learning algorithm is applied, considering the learning of a matrix of Q-values (Q-table) from the successive interactions with the environment.
The optimisation approach achieved very promising results by learning the optimal assembly sequence 98.3% of the times.
arXiv Detail & Related papers (2023-04-17T15:38:34Z) - Joint Training of Deep Ensembles Fails Due to Learner Collusion [61.557412796012535]
Ensembles of machine learning models have been well established as a powerful method of improving performance over a single model.
Traditionally, ensembling algorithms train their base learners independently or sequentially with the goal of optimizing their joint performance.
We show that directly minimizing the loss of the ensemble appears to rarely be applied in practice.
arXiv Detail & Related papers (2023-01-26T18:58:07Z) - Domain Adaptation Principal Component Analysis: base linear method for
learning with out-of-distribution data [55.41644538483948]
Domain adaptation is a popular paradigm in modern machine learning.
We present a method called Domain Adaptation Principal Component Analysis (DAPCA)
DAPCA finds a linear reduced data representation useful for solving the domain adaptation task.
arXiv Detail & Related papers (2022-08-28T21:10:56Z) - On the Convergence of Distributed Stochastic Bilevel Optimization
Algorithms over a Network [55.56019538079826]
Bilevel optimization has been applied to a wide variety of machine learning models.
Most existing algorithms restrict their single-machine setting so that they are incapable of handling distributed data.
We develop novel decentralized bilevel optimization algorithms based on a gradient tracking communication mechanism and two different gradients.
arXiv Detail & Related papers (2022-06-30T05:29:52Z) - DRFLM: Distributionally Robust Federated Learning with Inter-client
Noise via Local Mixup [58.894901088797376]
federated learning has emerged as a promising approach for training a global model using data from multiple organizations without leaking their raw data.
We propose a general framework to solve the above two challenges simultaneously.
We provide comprehensive theoretical analysis including robustness analysis, convergence analysis, and generalization ability.
arXiv Detail & Related papers (2022-04-16T08:08:29Z) - Multi-model Ensemble Learning Method for Human Expression Recognition [31.76775306959038]
We propose our solution based on the ensemble learning method to capture large amounts of real-life data.
We conduct many experiments on the AffWild2 dataset of the ABAW2022 Challenge, and the results demonstrate the effectiveness of our solution.
arXiv Detail & Related papers (2022-03-28T03:15:06Z) - Adaptive Sampling for Minimax Fair Classification [40.936345085421955]
We propose an adaptive sampling algorithm based on the principle of optimism, and derive theoretical bounds on its performance.
By deriving algorithm independent lower-bounds for a specific class of problems, we show that the performance achieved by our adaptive scheme cannot be improved in general.
arXiv Detail & Related papers (2021-03-01T04:58:27Z) - Small-Group Learning, with Application to Neural Architecture Search [17.86826990290058]
In human learning, a small group of students work together towards the same learning objective, where they express their understanding of a topic to their peers, compare their ideas, and help each other to trouble-shoot problems.
In this paper, we aim to investigate whether this human learning method can be borrowed to train better machine learning models, by developing a novel ML framework -- small-group learning (SGL)
SGL is formulated as a multi-level optimization framework consisting of three learning stages: each learner trains a model independently and uses this model to perform pseudo-labeling; each learner trains another model using datasets pseudo-
arXiv Detail & Related papers (2020-12-23T05:56:47Z) - Learning Diverse Representations for Fast Adaptation to Distribution
Shift [78.83747601814669]
We present a method for learning multiple models, incorporating an objective that pressures each to learn a distinct way to solve the task.
We demonstrate our framework's ability to facilitate rapid adaptation to distribution shift.
arXiv Detail & Related papers (2020-06-12T12:23:50Z) - Dynamic Federated Learning [57.14673504239551]
Federated learning has emerged as an umbrella term for centralized coordination strategies in multi-agent environments.
We consider a federated learning model where at every iteration, a random subset of available agents perform local updates based on their data.
Under a non-stationary random walk model on the true minimizer for the aggregate optimization problem, we establish that the performance of the architecture is determined by three factors, namely, the data variability at each agent, the model variability across all agents, and a tracking term that is inversely proportional to the learning rate of the algorithm.
arXiv Detail & Related papers (2020-02-20T15:00:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.