Fair Generative Models via Transfer Learning
- URL: http://arxiv.org/abs/2212.00926v1
- Date: Fri, 2 Dec 2022 01:44:38 GMT
- Title: Fair Generative Models via Transfer Learning
- Authors: Christopher TH Teo, Milad Abdollahzadeh, Ngai-Man Cheung
- Abstract summary: We propose fairTL, a transfer learning approach to learn fair generative models.
We introduce two additional innovations to improve upon fairTL: (i) multiple feedback and (ii) Linear-Probing followed by Fine-Tuning.
Extensive experiments show that fairTL and fairTL++ achieve state-of-the-art in both quality and fairness of generated samples.
- Score: 39.12323728810492
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This work addresses fair generative models. Dataset biases have been a major
cause of unfairness in deep generative models. Previous work had proposed to
augment large, biased datasets with small, unbiased reference datasets. Under
this setup, a weakly-supervised approach has been proposed, which achieves
state-of-the-art quality and fairness in generated samples. In our work, based
on this setup, we propose a simple yet effective approach. Specifically, first,
we propose fairTL, a transfer learning approach to learn fair generative
models. Under fairTL, we pre-train the generative model with the available
large, biased datasets and subsequently adapt the model using the small,
unbiased reference dataset. We find that our fairTL can learn expressive sample
generation during pre-training, thanks to the large (biased) dataset. This
knowledge is then transferred to the target model during adaptation, which also
learns to capture the underlying fair distribution of the small reference
dataset. Second, we propose fairTL++, where we introduce two additional
innovations to improve upon fairTL: (i) multiple feedback and (ii)
Linear-Probing followed by Fine-Tuning (LP-FT). Taking one step further, we
consider an alternative, challenging setup when only a pre-trained (potentially
biased) model is available but the dataset that was used to pre-train the model
is inaccessible. We demonstrate that our proposed fairTL and fairTL++ remain
very effective under this setup. We note that previous work requires access to
the large, biased datasets and is incapable of handling this more challenging
setup. Extensive experiments show that fairTL and fairTL++ achieve
state-of-the-art in both quality and fairness of generated samples. The code
and additional resources can be found at bearwithchris.github.io/fairTL/.
Related papers
- Dataset Ownership Verification in Contrastive Pre-trained Models [37.03747798645621]
We propose the first dataset ownership verification method tailored specifically for self-supervised pre-trained models by contrastive learning.
We validate the efficacy of this approach across multiple contrastive pre-trained models including SimCLR, BYOL, SimSiam, MOCO v3, and DINO.
arXiv Detail & Related papers (2025-02-11T05:42:21Z) - Universality in Transfer Learning for Linear Models [18.427215139020625]
We study the problem of transfer learning in linear models for both regression and binary classification.
We provide an exact and rigorous analysis and relate generalization errors (in regression) and classification errors (in binary classification) for the pretrained and fine-tuned models.
arXiv Detail & Related papers (2024-10-03T03:09:09Z) - Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review [50.78587571704713]
Learn-Focus-Review (LFR) is a dynamic training approach that adapts to the model's learning progress.
LFR tracks the model's learning performance across data blocks (sequences of tokens) and prioritizes revisiting challenging regions of the dataset.
Compared to baseline models trained on the full datasets, LFR consistently achieved lower perplexity and higher accuracy.
arXiv Detail & Related papers (2024-09-10T00:59:18Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Fair GANs through model rebalancing for extremely imbalanced class
distributions [5.463417677777276]
We present an approach to construct an unbiased generative adversarial network (GAN) from an existing biased GAN.
We show results for the StyleGAN2 models while training on the Flickr Faces High Quality (FFHQ) dataset for racial fairness.
We further validate our approach by applying it to an imbalanced CIFAR10 dataset which is also twice as large.
arXiv Detail & Related papers (2023-08-16T19:20:06Z) - Non-Invasive Fairness in Learning through the Lens of Data Drift [88.37640805363317]
We show how to improve the fairness of Machine Learning models without altering the data or the learning algorithm.
We use a simple but key insight: the divergence of trends between different populations, and, consecutively, between a learned model and minority populations, is analogous to data drift.
We explore two strategies (model-splitting and reweighing) to resolve this drift, aiming to improve the overall conformance of models to the underlying data.
arXiv Detail & Related papers (2023-03-30T17:30:42Z) - Beyond Transfer Learning: Co-finetuning for Action Localisation [64.07196901012153]
We propose co-finetuning -- simultaneously training a single model on multiple upstream'' and downstream'' tasks.
We demonstrate that co-finetuning outperforms traditional transfer learning when using the same total amount of data.
We also show how we can easily extend our approach to multiple upstream'' datasets to further improve performance.
arXiv Detail & Related papers (2022-07-08T10:25:47Z) - FairIF: Boosting Fairness in Deep Learning via Influence Functions with
Validation Set Sensitive Attributes [51.02407217197623]
We propose a two-stage training algorithm named FAIRIF.
It minimizes the loss over the reweighted data set where the sample weights are computed.
We show that FAIRIF yields models with better fairness-utility trade-offs against various types of bias.
arXiv Detail & Related papers (2022-01-15T05:14:48Z) - BiFair: Training Fair Models with Bilevel Optimization [8.2509884277533]
We develop a new training algorithm, named BiFair, which jointly minimizes for a utility, and a fairness loss of interest.
Our algorithm consistently performs better, i.e., we reach to better values of a given fairness metric under same, or higher accuracy.
arXiv Detail & Related papers (2021-06-03T22:36:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.