Domain Generalization via Rationale Invariance
- URL: http://arxiv.org/abs/2308.11158v1
- Date: Tue, 22 Aug 2023 03:31:40 GMT
- Title: Domain Generalization via Rationale Invariance
- Authors: Liang Chen, Yong Zhang, Yibing Song, Anton van den Hengel, and
Lingqiao Liu
- Abstract summary: This paper offers a new perspective to ease the challenge of domain generalization, which involves maintaining robust results even in unseen environments.
We propose treating the element-wise contributions to the final results as the rationale for making a decision and representing the rationale for each sample as a matrix.
Our experiments demonstrate that the proposed approach achieves competitive results across various datasets, despite its simplicity.
- Score: 70.32415695574555
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper offers a new perspective to ease the challenge of domain
generalization, which involves maintaining robust results even in unseen
environments. Our design focuses on the decision-making process in the final
classifier layer. Specifically, we propose treating the element-wise
contributions to the final results as the rationale for making a decision and
representing the rationale for each sample as a matrix. For a well-generalized
model, we suggest the rationale matrices for samples belonging to the same
category should be similar, indicating the model relies on domain-invariant
clues to make decisions, thereby ensuring robust results. To implement this
idea, we introduce a rationale invariance loss as a simple regularization
technique, requiring only a few lines of code. Our experiments demonstrate that
the proposed approach achieves competitive results across various datasets,
despite its simplicity. Code is available at
\url{https://github.com/liangchen527/RIDG}.
Related papers
- Obtaining Explainable Classification Models using Distributionally
Robust Optimization [12.511155426574563]
We study generalized linear models constructed using sets of feature value rules.
An inherent trade-off exists between rule set sparsity and its prediction accuracy.
We propose a new formulation to learn an ensemble of rule sets that simultaneously addresses these competing factors.
arXiv Detail & Related papers (2023-11-03T15:45:34Z) - Learning to Branch in Combinatorial Optimization with Graph Pointer
Networks [17.729352126574902]
This paper proposes a graph pointer network model for learning the variable selection policy in the branch-and-bound.
The proposed model, which combines the graph neural network and the pointer mechanism, can effectively map from the solver state to the branching variable decisions.
arXiv Detail & Related papers (2023-07-04T01:56:07Z) - Best-Effort Adaptation [62.00856290846247]
We present a new theoretical analysis of sample reweighting methods, including bounds holding uniformly over the weights.
We show how these bounds can guide the design of learning algorithms that we discuss in detail.
We report the results of a series of experiments demonstrating the effectiveness of our best-effort adaptation and domain adaptation algorithms.
arXiv Detail & Related papers (2023-05-10T00:09:07Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Predicting Out-of-Domain Generalization with Neighborhood Invariance [59.05399533508682]
We propose a measure of a classifier's output invariance in a local transformation neighborhood.
Our measure is simple to calculate, does not depend on the test point's true label, and can be applied even in out-of-domain (OOD) settings.
In experiments on benchmarks in image classification, sentiment analysis, and natural language inference, we demonstrate a strong and robust correlation between our measure and actual OOD generalization.
arXiv Detail & Related papers (2022-07-05T14:55:16Z) - Data-driven Numerical Invariant Synthesis with Automatic Generation of
Attributes [0.0]
We propose a data-driven algorithm for numerical invariant synthesis and verification.
The algorithm is based on the ICE-DT schema for learning decision trees from samples of positive and negative states.
arXiv Detail & Related papers (2022-05-30T09:18:37Z) - Attentional Prototype Inference for Few-Shot Segmentation [128.45753577331422]
We propose attentional prototype inference (API), a probabilistic latent variable framework for few-shot segmentation.
We define a global latent variable to represent the prototype of each object category, which we model as a probabilistic distribution.
We conduct extensive experiments on four benchmarks, where our proposal obtains at least competitive and often better performance than state-of-the-art prototype-based methods.
arXiv Detail & Related papers (2021-05-14T06:58:44Z) - SelfReg: Self-supervised Contrastive Regularization for Domain
Generalization [7.512471799525974]
We propose a new regularization method for domain generalization based on contrastive learning, self-supervised contrastive regularization (SelfReg)
The proposed approach use only positive data pairs, thus it resolves various problems caused by negative pair sampling.
In the recent benchmark, DomainBed, the proposed method shows comparable performance to the conventional state-of-the-art alternatives.
arXiv Detail & Related papers (2021-04-20T09:08:29Z) - Refined bounds for randomized experimental design [7.899055512130904]
Experimental design is an approach for selecting samples among a given set so as to obtain the best estimator for a given criterion.
We propose theoretical guarantees for randomized strategies on E and G-optimal design.
arXiv Detail & Related papers (2020-12-22T20:37:57Z) - Learning explanations that are hard to vary [75.30552491694066]
We show that averaging across examples can favor memorization and patchwork' solutions that sew together different strategies.
We then propose and experimentally validate a simple alternative algorithm based on a logical AND.
arXiv Detail & Related papers (2020-09-01T10:17:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.