RoCourseNet: Distributionally Robust Training of a Prediction Aware
Recourse Model
- URL: http://arxiv.org/abs/2206.00700v2
- Date: Fri, 18 Aug 2023 14:31:08 GMT
- Title: RoCourseNet: Distributionally Robust Training of a Prediction Aware
Recourse Model
- Authors: Hangzhi Guo, Feiran Jia, Jinghui Chen, Anna Squicciarini, Amulya Yadav
- Abstract summary: RoCourseNet is a training framework that jointly optimize predictions and recourses that are robust to future data shifts.
We show that RoCourseNet consistently achieves more than 96% robust validity and outperforms state-of-the-art baselines by at least 10% in generating robust explanations.
- Score: 29.057300578765663
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual (CF) explanations for machine learning (ML) models are
preferred by end-users, as they explain the predictions of ML models by
providing a recourse (or contrastive) case to individuals who are adversely
impacted by predicted outcomes. Existing CF explanation methods generate
recourses under the assumption that the underlying target ML model remains
stationary over time. However, due to commonly occurring distributional shifts
in training data, ML models constantly get updated in practice, which might
render previously generated recourses invalid and diminish end-users trust in
our algorithmic framework. To address this problem, we propose RoCourseNet, a
training framework that jointly optimizes predictions and recourses that are
robust to future data shifts. This work contains four key contributions: (1) We
formulate the robust recourse generation problem as a tri-level optimization
problem which consists of two sub-problems: (i) a bi-level problem that finds
the worst-case adversarial shift in the training data, and (ii) an outer
minimization problem to generate robust recourses against this worst-case
shift. (2) We leverage adversarial training to solve this tri-level
optimization problem by: (i) proposing a novel virtual data shift (VDS)
algorithm to find worst-case shifted ML models via explicitly considering the
worst-case data shift in the training dataset, and (ii) a block-wise coordinate
descent procedure to optimize for prediction and corresponding robust
recourses. (3) We evaluate RoCourseNet's performance on three real-world
datasets, and show that RoCourseNet consistently achieves more than 96% robust
validity and outperforms state-of-the-art baselines by at least 10% in
generating robust CF explanations. (4) Finally, we generalize the RoCourseNet
framework to accommodate any parametric post-hoc methods for improving robust
validity.
Related papers
- Scaling Laws for Predicting Downstream Performance in LLMs [75.28559015477137]
This work focuses on the pre-training loss as a more-efficient metric for performance estimation.
We extend the power law analytical function to predict domain-specific pre-training loss based on FLOPs across data sources.
We employ a two-layer neural network to model the non-linear relationship between multiple domain-specific loss and downstream performance.
arXiv Detail & Related papers (2024-10-11T04:57:48Z) - Towards Stable Machine Learning Model Retraining via Slowly Varying Sequences [6.067007470552307]
We propose a methodology for finding sequences of machine learning models that are stable across retraining iterations.
We develop a mixed-integer optimization formulation that is guaranteed to recover optimal models.
Our method shows stronger stability than greedily trained models with a small, controllable sacrifice in predictive power.
arXiv Detail & Related papers (2024-03-28T22:45:38Z) - Adversarial Collaborative Filtering for Free [27.949683060138064]
Collaborative Filtering (CF) has been successfully used to help users discover the items of interest.
Existing methods suffer from noisy data issue, which negatively impacts the quality of recommendation.
We present Sharpness-aware Collaborative Filtering (CF), a simple yet effective method that conducts adversarial training without extra computational cost over the base.
arXiv Detail & Related papers (2023-08-20T19:25:38Z) - DF2: Distribution-Free Decision-Focused Learning [53.2476224456902]
Decision-focused learning (DFL) has recently emerged as a powerful approach for predictthen-optimize problems.
Existing end-to-end DFL methods are hindered by three significant bottlenecks: model error, sample average approximation error, and distribution-based parameterization of the expected objective.
We present DF2 -- the first textit-free decision-focused learning method explicitly designed to address these three bottlenecks.
arXiv Detail & Related papers (2023-08-11T00:44:46Z) - DRFLM: Distributionally Robust Federated Learning with Inter-client
Noise via Local Mixup [58.894901088797376]
federated learning has emerged as a promising approach for training a global model using data from multiple organizations without leaking their raw data.
We propose a general framework to solve the above two challenges simultaneously.
We provide comprehensive theoretical analysis including robustness analysis, convergence analysis, and generalization ability.
arXiv Detail & Related papers (2022-04-16T08:08:29Z) - RoMA: Robust Model Adaptation for Offline Model-based Optimization [115.02677045518692]
We consider the problem of searching an input maximizing a black-box objective function given a static dataset of input-output queries.
A popular approach to solving this problem is maintaining a proxy model that approximates the true objective function.
Here, the main challenge is how to avoid adversarially optimized inputs during the search.
arXiv Detail & Related papers (2021-10-27T05:37:12Z) - CounterNet: End-to-End Training of Prediction Aware Counterfactual
Explanations [12.313007847721215]
CounterNet is an end-to-end learning framework which integrates predictive model training and the generation of counterfactual (CF) explanations.
Unlike post-hoc methods, CounterNet enables the optimization of the CF explanation generation only once together with the predictive model.
Our experiments on multiple real-world datasets show that CounterNet generates high-quality predictions.
arXiv Detail & Related papers (2021-09-15T20:09:13Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.