Investigating and Simplifying Masking-based Saliency Methods for Model
Interpretability
- URL: http://arxiv.org/abs/2010.09750v1
- Date: Mon, 19 Oct 2020 18:00:36 GMT
- Title: Investigating and Simplifying Masking-based Saliency Methods for Model
Interpretability
- Authors: Jason Phang, Jungkyu Park and Krzysztof J. Geras
- Abstract summary: Saliency maps that identify the most informative regions of an image are valuable for model interpretability.
A common approach to creating saliency maps involves generating input masks that mask out portions of an image.
We show that a masking model can be trained with as few as 10 examples per class and still generate saliency maps with only a 0.7-point increase in localization error.
- Score: 5.387323728379395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Saliency maps that identify the most informative regions of an image for a
classifier are valuable for model interpretability. A common approach to
creating saliency maps involves generating input masks that mask out portions
of an image to maximally deteriorate classification performance, or mask in an
image to preserve classification performance. Many variants of this approach
have been proposed in the literature, such as counterfactual generation and
optimizing over a Gumbel-Softmax distribution. Using a general formulation of
masking-based saliency methods, we conduct an extensive evaluation study of a
number of recently proposed variants to understand which elements of these
methods meaningfully improve performance. Surprisingly, we find that a
well-tuned, relatively simple formulation of a masking-based saliency model
outperforms many more complex approaches. We find that the most important
ingredients for high quality saliency map generation are (1) using both
masked-in and masked-out objectives and (2) training the classifier alongside
the masking model. Strikingly, we show that a masking model can be trained with
as few as 10 examples per class and still generate saliency maps with only a
0.7-point increase in localization error.
Related papers
- Hybrid diffusion models: combining supervised and generative pretraining for label-efficient fine-tuning of segmentation models [55.2480439325792]
We propose a new pretext task, which is to perform simultaneously image denoising and mask prediction on the first domain.
We show that fine-tuning a model pretrained using this approach leads to better results than fine-tuning a similar model trained using either supervised or unsupervised pretraining.
arXiv Detail & Related papers (2024-08-06T20:19:06Z) - MaskInversion: Localized Embeddings via Optimization of Explainability Maps [49.50785637749757]
MaskInversion generates a context-aware embedding for a query image region specified by a mask at test time.
It can be used for a broad range of tasks, including open-vocabulary class retrieval, referring expression comprehension, as well as for localized captioning and image generation.
arXiv Detail & Related papers (2024-07-29T14:21:07Z) - Mask2Anomaly: Mask Transformer for Universal Open-set Segmentation [29.43462426812185]
We propose a paradigm change by shifting from a per-pixel classification to a mask classification.
Our mask-based method, Mask2Anomaly, demonstrates the feasibility of integrating a mask-classification architecture.
By comprehensive qualitative and qualitative evaluation, we show Mask2Anomaly achieves new state-of-the-art results.
arXiv Detail & Related papers (2023-09-08T20:07:18Z) - Unmasking Anomalies in Road-Scene Segmentation [18.253109627901566]
Anomaly segmentation is a critical task for driving applications.
We propose a paradigm change by shifting from a per-pixel classification to a mask classification.
Mask2Anomaly demonstrates the feasibility of integrating an anomaly detection method in a mask-classification architecture.
arXiv Detail & Related papers (2023-07-25T08:23:10Z) - Denoising Diffusion Semantic Segmentation with Mask Prior Modeling [61.73352242029671]
We propose to ameliorate the semantic segmentation quality of existing discriminative approaches with a mask prior modeled by a denoising diffusion generative model.
We evaluate the proposed prior modeling with several off-the-shelf segmentors, and our experimental results on ADE20K and Cityscapes demonstrate that our approach could achieve competitively quantitative performance.
arXiv Detail & Related papers (2023-06-02T17:47:01Z) - Improving Masked Autoencoders by Learning Where to Mask [65.89510231743692]
Masked image modeling is a promising self-supervised learning method for visual data.
We present AutoMAE, a framework that uses Gumbel-Softmax to interlink an adversarially-trained mask generator and a mask-guided image modeling process.
In our experiments, AutoMAE is shown to provide effective pretraining models on standard self-supervised benchmarks and downstream tasks.
arXiv Detail & Related papers (2023-03-12T05:28:55Z) - What You See is What You Classify: Black Box Attributions [61.998683569022006]
We train a deep network, the Explainer, to predict attributions for a pre-trained black-box classifier, the Explanandum.
Unlike most existing approaches, ours is capable of directly generating very distinct class-specific masks.
We show that our attributions are superior to established methods both visually and quantitatively.
arXiv Detail & Related papers (2022-05-23T12:30:04Z) - ContrastMask: Contrastive Learning to Segment Every Thing [18.265503138997794]
We propose ContrastMask, which learns a mask segmentation model on both seen and unseen categories.
Features from the mask regions (foreground) are pulled together, and are contrasted against those from the background, and vice versa.
Exhaustive experiments on the COCO dataset demonstrate the superiority of our method.
arXiv Detail & Related papers (2022-03-18T07:41:48Z) - Per-Pixel Classification is Not All You Need for Semantic Segmentation [184.2905747595058]
Mask classification is sufficiently general to solve both semantic- and instance-level segmentation tasks.
We propose MaskFormer, a simple mask classification model which predicts a set of binary masks.
Our method outperforms both current state-of-the-art semantic (55.6 mIoU on ADE20K) and panoptic segmentation (52.7 PQ on COCO) models.
arXiv Detail & Related papers (2021-07-13T17:59:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.