Dual Decomposition of Convex Optimization Layers for Consistent
Attention in Medical Images
- URL: http://arxiv.org/abs/2206.02761v2
- Date: Tue, 7 Jun 2022 14:07:51 GMT
- Title: Dual Decomposition of Convex Optimization Layers for Consistent
Attention in Medical Images
- Authors: Tom Ron, Michal Weiler-Sagie, Tamir Hazan
- Abstract summary: Key concern in integrating machine learning models in medicine is the ability to interpret their reasoning.
We propose a multi-layer attention mechanism that enforces consistent interpretations between attended convolutional layers.
Our proposed method leverages weakly annotated medical imaging data and provides complete and faithful explanations to the model's prediction.
- Score: 12.844658658362325
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A key concern in integrating machine learning models in medicine is the
ability to interpret their reasoning. Popular explainability methods have
demonstrated satisfactory results in natural image recognition, yet in medical
image analysis, many of these approaches provide partial and noisy
explanations. Recently, attention mechanisms have shown compelling results both
in their predictive performance and in their interpretable qualities. A
fundamental trait of attention is that it leverages salient parts of the input
which contribute to the model's prediction. To this end, our work focuses on
the explanatory value of attention weight distributions. We propose a
multi-layer attention mechanism that enforces consistent interpretations
between attended convolutional layers using convex optimization. We apply
duality to decompose the consistency constraints between the layers by
reparameterizing their attention probability distributions. We further suggest
learning the dual witness by optimizing with respect to our objective; thus,
our implementation uses standard back-propagation, hence it is highly
efficient. While preserving predictive performance, our proposed method
leverages weakly annotated medical imaging data and provides complete and
faithful explanations to the model's prediction.
Related papers
- Improving Explainability of Disentangled Representations using
Multipath-Attribution Mappings [12.145748796751619]
We propose a framework that utilizes interpretable disentangled representations for downstream-task prediction.
We demonstrate the effectiveness of our approach on a synthetic benchmark suite and two medical datasets.
arXiv Detail & Related papers (2023-06-15T10:52:29Z) - Two-step interpretable modeling of Intensive Care Acquired Infections [0.0]
We present a novel methodology for integrating high resolution longitudinal data with the dynamic prediction capabilities of survival models.
The aim is two-fold: to improve the predictive power while maintaining interpretability of the models.
arXiv Detail & Related papers (2023-01-26T14:54:17Z) - Hybrid Predictive Coding: Inferring, Fast and Slow [62.997667081978825]
We propose a hybrid predictive coding network that combines both iterative and amortized inference in a principled manner.
We demonstrate that our model is inherently sensitive to its uncertainty and adaptively balances balances to obtain accurate beliefs using minimum computational expense.
arXiv Detail & Related papers (2022-04-05T12:52:45Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Proactive Pseudo-Intervention: Causally Informed Contrastive Learning
For Interpretable Vision Models [103.64435911083432]
We present a novel contrastive learning strategy called it Proactive Pseudo-Intervention (PPI)
PPI leverages proactive interventions to guard against image features with no causal relevance.
We also devise a novel causally informed salience mapping module to identify key image pixels to intervene, and show it greatly facilitates model interpretability.
arXiv Detail & Related papers (2020-12-06T20:30:26Z) - Double Robust Representation Learning for Counterfactual Prediction [68.78210173955001]
We propose a novel scalable method to learn double-robust representations for counterfactual predictions.
We make robust and efficient counterfactual predictions for both individual and average treatment effects.
The algorithm shows competitive performance with the state-of-the-art on real world and synthetic data.
arXiv Detail & Related papers (2020-10-15T16:39:26Z) - Domain aware medical image classifier interpretation by counterfactual
impact analysis [2.512212190779389]
We introduce a neural-network based attribution method, applicable to any trained predictor.
Our solution identifies salient regions of an input image in a single forward-pass by measuring the effect of local image-perturbations on a predictor's score.
arXiv Detail & Related papers (2020-07-13T11:11:17Z) - Explain and Improve: LRP-Inference Fine-Tuning for Image Captioning
Models [82.3793660091354]
This paper analyzes the predictions of image captioning models with attention mechanisms beyond visualizing the attention itself.
We develop variants of layer-wise relevance propagation (LRP) and gradient-based explanation methods, tailored to image captioning models with attention mechanisms.
arXiv Detail & Related papers (2020-01-04T05:15:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.