Model Transferability With Responsive Decision Subjects
- URL: http://arxiv.org/abs/2107.05911v4
- Date: Thu, 1 Jun 2023 16:07:11 GMT
- Title: Model Transferability With Responsive Decision Subjects
- Authors: Yatong Chen, Zeyu Tang, Kun Zhang, Yang Liu
- Abstract summary: We formalize the discussions of the transferability of a model by studying how the performance of the model trained on the available source distribution would translate to the performance on its induced domain.
We provide both upper bounds for the performance gap due to the induced domain shift, as well as lower bounds for the trade-offs that a classifier has to suffer on either the source training distribution or the induced target distribution.
- Score: 11.07759054787023
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Given an algorithmic predictor that is accurate on some source population
consisting of strategic human decision subjects, will it remain accurate if the
population respond to it? In our setting, an agent or a user corresponds to a
sample $(X,Y)$ drawn from a distribution $\cal{D}$ and will face a model $h$
and its classification result $h(X)$. Agents can modify $X$ to adapt to $h$,
which will incur a distribution shift on $(X,Y)$. Our formulation is motivated
by applications where the deployed machine learning models are subjected to
human agents, and will ultimately face responsive and interactive data
distributions. We formalize the discussions of the transferability of a model
by studying how the performance of the model trained on the available source
distribution (data) would translate to the performance on its induced domain.
We provide both upper bounds for the performance gap due to the induced domain
shift, as well as lower bounds for the trade-offs that a classifier has to
suffer on either the source training distribution or the induced target
distribution. We provide further instantiated analysis for two popular domain
adaptation settings, including covariate shift and target shift.
Related papers
- Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Fairness Hub Technical Briefs: Definition and Detection of Distribution Shift [0.5825410941577593]
Distribution shift is a common situation in machine learning tasks, where the data used for training a model is different from the data the model is applied to in the real world.
This brief focuses on the definition and detection of distribution shifts in educational settings.
arXiv Detail & Related papers (2024-05-23T05:29:36Z) - Delta-AI: Local objectives for amortized inference in sparse graphical models [64.5938437823851]
We present a new algorithm for amortized inference in sparse probabilistic graphical models (PGMs)
Our approach is based on the observation that when the sampling of variables in a PGM is seen as a sequence of actions taken by an agent, sparsity of the PGM enables local credit assignment in the agent's policy learning objective.
We illustrate $Delta$-AI's effectiveness for sampling from synthetic PGMs and training latent variable models with sparse factor structure.
arXiv Detail & Related papers (2023-10-03T20:37:03Z) - Diagnosing Model Performance Under Distribution Shift [9.143551270841858]
Prediction models can perform poorly when deployed to target distributions different from the training distribution.
Our approach decomposes the performance drop into terms for 1) an increase in harder but frequently seen examples from training, 2) changes in the relationship between features and outcomes, and 3) poor performance on examples infrequent or unseen during training.
arXiv Detail & Related papers (2023-03-03T15:27:16Z) - Explaining Cross-Domain Recognition with Interpretable Deep Classifier [100.63114424262234]
Interpretable Deep (IDC) learns the nearest source samples of a target sample as evidence upon which the classifier makes the decision.
Our IDC leads to a more explainable model with almost no accuracy degradation and effectively calibrates classification for optimum reject options.
arXiv Detail & Related papers (2022-11-15T15:58:56Z) - Domain Adaptation meets Individual Fairness. And they get along [48.95808607591299]
We show that algorithmic fairness interventions can help machine learning models overcome distribution shifts.
In particular, we show that enforcing suitable notions of individual fairness (IF) can improve the out-of-distribution accuracy of ML models.
arXiv Detail & Related papers (2022-05-01T16:19:55Z) - Predicting with Confidence on Unseen Distributions [90.68414180153897]
We connect domain adaptation and predictive uncertainty literature to predict model accuracy on challenging unseen distributions.
We find that the difference of confidences (DoC) of a classifier's predictions successfully estimates the classifier's performance change over a variety of shifts.
We specifically investigate the distinction between synthetic and natural distribution shifts and observe that despite its simplicity DoC consistently outperforms other quantifications of distributional difference.
arXiv Detail & Related papers (2021-07-07T15:50:18Z) - Estimating Generalization under Distribution Shifts via Domain-Invariant
Representations [75.74928159249225]
We use a set of domain-invariant predictors as a proxy for the unknown, true target labels.
The error of the resulting risk estimate depends on the target risk of the proxy model.
arXiv Detail & Related papers (2020-07-06T17:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.