On the explainable properties of 1-Lipschitz Neural Networks: An Optimal
Transport Perspective
- URL: http://arxiv.org/abs/2206.06854v3
- Date: Fri, 2 Feb 2024 08:59:26 GMT
- Title: On the explainable properties of 1-Lipschitz Neural Networks: An Optimal
Transport Perspective
- Authors: Mathieu Serrurier (IRIT-ADRIA, UT), Franck Mamalet (UT), Thomas Fel
(UT), Louis B\'ethune (UT3, UT, IRIT-ADRIA), Thibaut Boissin (UT)
- Abstract summary: Saliency Maps generated by traditional neural networks are often noisy and provide limited insights.
In this paper, we demonstrate that, on the contrary, the Saliency Maps of 1-Lipschitz neural networks, exhibit desirable XAI properties.
We also prove that these maps align unprecedentedly well with human explanations on ImageNet.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Input gradients have a pivotal role in a variety of applications, including
adversarial attack algorithms for evaluating model robustness, explainable AI
techniques for generating Saliency Maps, and counterfactual
explanations.However, Saliency Maps generated by traditional neural networks
are often noisy and provide limited insights. In this paper, we demonstrate
that, on the contrary, the Saliency Maps of 1-Lipschitz neural networks,
learned with the dual loss of an optimal transportation problem, exhibit
desirable XAI properties:They are highly concentrated on the essential parts of
the image with low noise, significantly outperforming state-of-the-art
explanation approaches across various models and metrics. We also prove that
these maps align unprecedentedly well with human explanations on ImageNet.To
explain the particularly beneficial properties of the Saliency Map for such
models, we prove this gradient encodes both the direction of the transportation
plan and the direction towards the nearest adversarial attack. Following the
gradient down to the decision boundary is no longer considered an adversarial
attack, but rather a counterfactual explanation that explicitly transports the
input from one class to another. Thus, Learning with such a loss jointly
optimizes the classification objective and the alignment of the gradient, i.e.
the Saliency Map, to the transportation plan direction.These networks were
previously known to be certifiably robust by design, and we demonstrate that
they scale well for large problems and models, and are tailored for
explainability using a fast and straightforward method.
Related papers
- TopoSD: Topology-Enhanced Lane Segment Perception with SDMap Prior [70.84644266024571]
We propose to train a perception model to "see" standard definition maps (SDMaps)
We encode SDMap elements into neural spatial map representations and instance tokens, and then incorporate such complementary features as prior information.
Based on the lane segment representation framework, the model simultaneously predicts lanes, centrelines and their topology.
arXiv Detail & Related papers (2024-11-22T06:13:42Z) - Landscaping Linear Mode Connectivity [76.39694196535996]
linear mode connectivity (LMC) has garnered interest from both theoretical and practical fronts.
We take a step towards understanding it by providing a model of how the loss landscape needs to behave topographically for LMC.
arXiv Detail & Related papers (2024-06-24T03:53:30Z) - SCAAT: Improving Neural Network Interpretability via Saliency
Constrained Adaptive Adversarial Training [10.716021768803433]
Saliency map is a common form of explanation illustrating the heatmap of feature attributions.
We propose a model-agnostic learning method called Saliency Constrained Adaptive Adversarial Training (SCAAT) to improve the quality of such DNN interpretability.
arXiv Detail & Related papers (2023-11-09T04:48:38Z) - MacFormer: Map-Agent Coupled Transformer for Real-time and Robust
Trajectory Prediction [26.231420111336565]
We propose Map-Agent Coupled Transformer (MacFormer) for real-time and robust trajectory prediction.
Our framework explicitly incorporates map constraints into the network via two carefully designed modules named coupled map and reference extractor.
We evaluate our approach on Argoverse 1, Argoverse 2, and nuScenes real-world benchmarks, where it all achieved state-of-the-art performance.
arXiv Detail & Related papers (2023-08-20T14:27:28Z) - Shap-CAM: Visual Explanations for Convolutional Neural Networks based on
Shapley Value [86.69600830581912]
We develop a novel visual explanation method called Shap-CAM based on class activation mapping.
We demonstrate that Shap-CAM achieves better visual performance and fairness for interpreting the decision making process.
arXiv Detail & Related papers (2022-08-07T00:59:23Z) - Unpaired Image Super-Resolution with Optimal Transport Maps [128.1189695209663]
Real-world image super-resolution (SR) tasks often do not have paired datasets limiting the application of supervised techniques.
We propose an algorithm for unpaired SR which learns an unbiased OT map for the perceptual transport cost.
Our algorithm provides nearly state-of-the-art performance on the large-scale unpaired AIM-19 dataset.
arXiv Detail & Related papers (2022-02-02T16:21:20Z) - CAMERAS: Enhanced Resolution And Sanity preserving Class Activation
Mapping for image saliency [61.40511574314069]
Backpropagation image saliency aims at explaining model predictions by estimating model-centric importance of individual pixels in the input.
We propose CAMERAS, a technique to compute high-fidelity backpropagation saliency maps without requiring any external priors.
arXiv Detail & Related papers (2021-06-20T08:20:56Z) - Assessing the Reliability of Visual Explanations of Deep Models with
Adversarial Perturbations [15.067369314723958]
We propose an objective measure to evaluate the reliability of explanations of deep models.
Our approach is based on changes in the network's outcome resulting from the perturbation of input images in an adversarial way.
We also propose a straightforward application of our approach to clean relevance maps, creating more interpretable maps without any loss in essential explanation.
arXiv Detail & Related papers (2020-04-22T19:57:34Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.