Exposing Image Classifier Shortcuts with Counterfactual Frequency (CoF) Tables
- URL: http://arxiv.org/abs/2405.15661v1
- Date: Fri, 24 May 2024 15:58:02 GMT
- Title: Exposing Image Classifier Shortcuts with Counterfactual Frequency (CoF) Tables
- Authors: James Hinns, David Martens,
- Abstract summary: 'Shortcuts' are easy-to-learn patterns from the training data that fail to generalise to new data.
Examples include the use of a copyright watermark to recognise horses, snowy background to recognise huskies, or ink markings to detect malignant skin lesions.
We introduce Counterfactual Frequency tables, a novel approach that aggregates instance-based explanations into global insights.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The rise of deep learning in image classification has brought unprecedented accuracy but also highlighted a key issue: the use of 'shortcuts' by models. Such shortcuts are easy-to-learn patterns from the training data that fail to generalise to new data. Examples include the use of a copyright watermark to recognise horses, snowy background to recognise huskies, or ink markings to detect malignant skin lesions. The explainable AI (XAI) community has suggested using instance-level explanations to detect shortcuts without external data, but this requires the examination of many explanations to confirm the presence of such shortcuts, making it a labour-intensive process. To address these challenges, we introduce Counterfactual Frequency (CoF) tables, a novel approach that aggregates instance-based explanations into global insights, and exposes shortcuts. The aggregation implies the need for some semantic concepts to be used in the explanations, which we solve by labelling the segments of an image. We demonstrate the utility of CoF tables across several datasets, revealing the shortcuts learned from them.
Related papers
- A Good Foundation is Worth Many Labels: Label-Efficient Panoptic Segmentation [22.440065488051047]
Key challenge for the widespread application of learning-based models for robotic perception is to significantly reduce the required amount of annotated training data.
We exploit the groundwork paved by visual foundation models to train two lightweight network heads for semantic segmentation and object boundary detection.
We demonstrate that PASTEL significantly outperforms previous methods for label-efficient segmentation even when using fewer annotations.
arXiv Detail & Related papers (2024-05-29T12:23:29Z) - How to Construct Perfect and Worse-than-Coin-Flip Spoofing
Countermeasures: A Word of Warning on Shortcut Learning [20.486639064376014]
Shortcut learning, or Clever Hans effect refers to situations where a learning agent learns spurious correlations present in data, resulting in biased models.
We focus on finding shortcuts in deep learning based spoofing countermeasures (CMs) that predict whether a given utterance is spoofed or not.
arXiv Detail & Related papers (2023-05-31T15:58:37Z) - Neglected Free Lunch -- Learning Image Classifiers Using Annotation
Byproducts [43.76258241948858]
Supervised learning of image classifiers distills human knowledge into a parametric model through pairs of images and corresponding labels (X,Y)
We argue that this simple and widely used representation of human knowledge neglects rich auxiliary information from the annotation procedure.
We refer to the new paradigm of training models with annotation byproducts as learning using annotation byproducts (LUAB)
arXiv Detail & Related papers (2023-03-30T17:59:02Z) - Shortcut Detection with Variational Autoencoders [1.3174512123890016]
We present a novel approach to detect shortcuts in image and audio datasets by leveraging variational autoencoders (VAEs)
The disentanglement of features in the latent space of VAEs allows us to discover feature-target correlations in datasets and semi-automatically evaluate them for ML shortcuts.
We demonstrate the applicability of our method on several real-world datasets and identify shortcuts that have not been discovered before.
arXiv Detail & Related papers (2023-02-08T18:26:10Z) - Which Shortcut Solution Do Question Answering Models Prefer to Learn? [38.36299280464046]
Question answering (QA) models for reading comprehension tend to learn shortcut solutions rather than the solutions intended by QA datasets.
We show that shortcuts that exploit answer positions and word-label correlations are preferentially learned for extractive and multiple-choice QA.
We experimentally show that the learnability of shortcuts can be utilized to construct an effective QA training set.
arXiv Detail & Related papers (2022-11-29T13:57:59Z) - No Token Left Behind: Explainability-Aided Image Classification and
Generation [79.4957965474334]
We present a novel explainability-based approach, which adds a loss term to ensure that CLIP focuses on all relevant semantic parts of the input.
Our method yields an improvement in the recognition rate, without additional training or fine-tuning.
arXiv Detail & Related papers (2022-04-11T07:16:39Z) - Weakly Supervised Semantic Segmentation using Out-of-Distribution Data [50.45689349004041]
Weakly supervised semantic segmentation (WSSS) methods are often built on pixel-level localization maps.
We propose a novel source of information to distinguish foreground from the background: Out-of-Distribution (OoD) data.
arXiv Detail & Related papers (2022-03-08T05:33:35Z) - Learning to Detect Instance-level Salient Objects Using Complementary
Image Labels [55.049347205603304]
We present the first weakly-supervised approach to the salient instance detection problem.
We propose a novel weakly-supervised network with three branches: a Saliency Detection Branch leveraging class consistency information to locate candidate objects; a Boundary Detection Branch exploiting class discrepancy information to delineate object boundaries; and a Centroid Detection Branch using subitizing information to detect salient instance centroids.
arXiv Detail & Related papers (2021-11-19T10:15:22Z) - Why Machine Reading Comprehension Models Learn Shortcuts? [56.629192589376046]
We argue that larger proportion of shortcut questions in training data make models rely on shortcut tricks excessively.
A thorough empirical analysis shows that MRC models tend to learn shortcut questions earlier than challenging questions.
arXiv Detail & Related papers (2021-06-02T08:43:12Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z) - Region Comparison Network for Interpretable Few-shot Image
Classification [97.97902360117368]
Few-shot image classification has been proposed to effectively use only a limited number of labeled examples to train models for new classes.
We propose a metric learning based method named Region Comparison Network (RCN), which is able to reveal how few-shot learning works.
We also present a new way to generalize the interpretability from the level of tasks to categories.
arXiv Detail & Related papers (2020-09-08T07:29:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.