Beyond Discriminant Patterns: On the Robustness of Decision Rule
Ensembles
- URL: http://arxiv.org/abs/2109.10432v1
- Date: Tue, 21 Sep 2021 20:50:10 GMT
- Title: Beyond Discriminant Patterns: On the Robustness of Decision Rule
Ensembles
- Authors: Xin Du, Subramanian Ramamoorthy, Wouter Duivesteijn, Jin Tian, Mykola
Pechenizkiy
- Abstract summary: Local decision rules are commonly understood to be more explainable, due to the local nature of the patterns involved.
We propose a new method to learn and ensemble local decision rules, that are robust both in the training and deployment environments.
- Score: 22.57678894050416
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Local decision rules are commonly understood to be more explainable, due to
the local nature of the patterns involved. With numerical optimization methods
such as gradient boosting, ensembles of local decision rules can gain good
predictive performance on data involving global structure. Meanwhile, machine
learning models are being increasingly used to solve problems in high-stake
domains including healthcare and finance. Here, there is an emerging consensus
regarding the need for practitioners to understand whether and how those models
could perform robustly in the deployment environments, in the presence of
distributional shifts. Past research on local decision rules has focused mainly
on maximizing discriminant patterns, without due consideration of robustness
against distributional shifts. In order to fill this gap, we propose a new
method to learn and ensemble local decision rules, that are robust both in the
training and deployment environments. Specifically, we propose to leverage
causal knowledge by regarding the distributional shifts in subpopulations and
deployment environments as the results of interventions on the underlying
system. We propose two regularization terms based on causal knowledge to search
for optimal and stable rules. Experiments on both synthetic and benchmark
datasets show that our method is effective and robust against distributional
shifts in multiple environments.
Related papers
- Decision-Focused Evaluation of Worst-Case Distribution Shift [18.98504221245623]
We introduce a novel framework to identify worst-case distribution shifts in predictive resource allocation settings.
We show that the problem can be reformulated as a submodular optimization problem, enabling efficient approximations of worst-case loss.
Applying our framework to real data, we find empirical evidence that worst-case shifts identified by one metric often significantly diverge from worst-case distributions identified by other metrics.
arXiv Detail & Related papers (2024-07-04T01:00:53Z) - Differentiable Distributionally Robust Optimization Layers [10.667165962654996]
We develop differentiable DRO layers for generic mixed-integer DRO problems with parameterized second-order conic ambiguity sets.
We propose a novel dual-view methodology by handling continuous and discrete parts of decisions via different principles.
Specifically, we construct a differentiable energy-based surrogate to implement the dual-view methodology and use importance sampling to estimate its gradient.
arXiv Detail & Related papers (2024-06-24T12:09:19Z) - Proxy Methods for Domain Adaptation [78.03254010884783]
proxy variables allow for adaptation to distribution shift without explicitly recovering or modeling latent variables.
We develop a two-stage kernel estimation approach to adapt to complex distribution shifts in both settings.
arXiv Detail & Related papers (2024-03-12T09:32:41Z) - Adaptive Global-Local Representation Learning and Selection for
Cross-Domain Facial Expression Recognition [54.334773598942775]
Domain shift poses a significant challenge in Cross-Domain Facial Expression Recognition (CD-FER)
We propose an Adaptive Global-Local Representation Learning and Selection framework.
arXiv Detail & Related papers (2024-01-20T02:21:41Z) - Robust Networked Federated Learning for Localization [7.332862402432447]
This paper addresses the problem of approximation, which is non-smooth in a federated setting where the data is distributed across a multitude of devices.
We propose a method that adopts an $L_$-norm that a robust formulation within a distributed subgradient framework, explicitly designed to handle these obstacles.
arXiv Detail & Related papers (2023-08-31T13:54:37Z) - Statistical Inference Under Constrained Selection Bias [20.862583584531322]
We propose a framework that enables statistical inference in the presence of selection bias.
The output is high-probability bounds on the value of an estimand for the target distribution.
We analyze the computational and statistical properties of methods to estimate these bounds and show that our method can produce informative bounds on a variety of simulated and semisynthetic tasks.
arXiv Detail & Related papers (2023-06-05T23:05:26Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities [125.62877849447729]
We consider distributed variational inequalities (VIs) on domains with the problem data that is heterogeneous (non-IID) and distributed across many devices.
We make a very general assumption on the computational network that covers the settings of fully decentralized calculations.
We theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone settings.
arXiv Detail & Related papers (2021-06-15T17:45:51Z) - Robust Generalization despite Distribution Shift via Minimum
Discriminating Information [46.164498176119665]
We introduce a modeling framework where, in addition to training data, we have partial structural knowledge of the shifted test distribution.
We employ the principle of minimum discriminating information to embed the available prior knowledge.
We obtain explicit generalization bounds with respect to the unknown shifted distribution.
arXiv Detail & Related papers (2021-06-08T15:25:35Z) - Learning Calibrated Uncertainties for Domain Shift: A Distributionally
Robust Learning Approach [150.8920602230832]
We propose a framework for learning calibrated uncertainties under domain shifts.
In particular, the density ratio estimation reflects the closeness of a target (test) sample to the source (training) distribution.
We show that our proposed method generates calibrated uncertainties that benefit downstream tasks.
arXiv Detail & Related papers (2020-10-08T02:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.