Explainable Multi-View Deep Networks Methodology for Experimental
Physics
- URL: http://arxiv.org/abs/2308.08206v2
- Date: Thu, 17 Aug 2023 04:58:55 GMT
- Title: Explainable Multi-View Deep Networks Methodology for Experimental
Physics
- Authors: Nadav Schneider, Muriel Tzdaka, Galit Sturm, Guy Lazovski, Galit Bar,
Gilad Oren, Raz Gvishi, Gal Oren
- Abstract summary: Physical experiments often involve multiple imaging representations, such as X-ray scans and microscopic images.
Deep learning models have been widely used for supervised analysis in these experiments.
Multi-view data has emerged, where each sample is described by views from different angles, sources, or modalities.
There is a lack of proper explainability in multi-view models, which are challenging to explain due to their architectures.
- Score: 0.20417058495510368
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Physical experiments often involve multiple imaging representations, such as
X-ray scans and microscopic images. Deep learning models have been widely used
for supervised analysis in these experiments. Combining different image
representations is frequently required to analyze and make a decision properly.
Consequently, multi-view data has emerged - datasets where each sample is
described by views from different angles, sources, or modalities. These
problems are addressed with the concept of multi-view learning. Understanding
the decision-making process of deep learning models is essential for reliable
and credible analysis. Hence, many explainability methods have been devised
recently. Nonetheless, there is a lack of proper explainability in multi-view
models, which are challenging to explain due to their architectures. In this
paper, we suggest different multi-view architectures for the vision domain,
each suited to another problem, and we also present a methodology for
explaining these models. To demonstrate the effectiveness of our methodology,
we focus on the domain of High Energy Density Physics (HEDP) experiments, where
multiple imaging representations are used to assess the quality of foam
samples. We apply our methodology to classify the foam samples quality using
the suggested multi-view architectures. Through experimental results, we
showcase the improvement of accurate architecture choice on both accuracy - 78%
to 84% and AUC - 83% to 93% and present a trade-off between performance and
explainability. Specifically, we demonstrate that our approach enables the
explanation of individual one-view models, providing insights into the
decision-making process of each view. This understanding enhances the
interpretability of the overall multi-view model. The sources of this work are
available at:
https://github.com/Scientific-Computing-Lab-NRCN/Multi-View-Explainability.
Related papers
- Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics [54.08757792080732]
We propose integrating deep features from pre-trained visual models with a statistical analysis model to achieve opinion-unaware BIQA (OU-BIQA)
Our proposed model exhibits superior consistency with human visual perception compared to state-of-the-art BIQA models.
arXiv Detail & Related papers (2024-05-29T06:09:34Z) - Revealing Multimodal Contrastive Representation Learning through Latent
Partial Causal Models [85.67870425656368]
We introduce a unified causal model specifically designed for multimodal data.
We show that multimodal contrastive representation learning excels at identifying latent coupled variables.
Experiments demonstrate the robustness of our findings, even when the assumptions are violated.
arXiv Detail & Related papers (2024-02-09T07:18:06Z) - OtterHD: A High-Resolution Multi-modality Model [57.16481886807386]
OtterHD-8B is an innovative multimodal model engineered to interpret high-resolution visual inputs with granular precision.
Our study highlights the critical role of flexibility and high-resolution input capabilities in large multimodal models.
arXiv Detail & Related papers (2023-11-07T18:59:58Z) - StableLLaVA: Enhanced Visual Instruction Tuning with Synthesized
Image-Dialogue Data [129.92449761766025]
We propose a novel data collection methodology that synchronously synthesizes images and dialogues for visual instruction tuning.
This approach harnesses the power of generative models, marrying the abilities of ChatGPT and text-to-image generative models.
Our research includes comprehensive experiments conducted on various datasets.
arXiv Detail & Related papers (2023-08-20T12:43:52Z) - Domain Generalization for Mammographic Image Analysis with Contrastive
Learning [62.25104935889111]
The training of an efficacious deep learning model requires large data with diverse styles and qualities.
A novel contrastive learning is developed to equip the deep learning models with better style generalization capability.
The proposed method has been evaluated extensively and rigorously with mammograms from various vendor style domains and several public datasets.
arXiv Detail & Related papers (2023-04-20T11:40:21Z) - Multispectral Contrastive Learning with Viewmaker Networks [8.635434871127512]
We focus on applying contrastive learning approaches to a variety of remote sensing datasets.
We show that Viewmaker networks are promising for producing views in this setting without requiring extensive domain knowledge and trial and error.
arXiv Detail & Related papers (2023-02-11T18:44:12Z) - Encoding Domain Knowledge in Multi-view Latent Variable Models: A
Bayesian Approach with Structured Sparsity [7.811916700683125]
MuVI is a novel approach for domain-informed multi-view latent variable models.
We demonstrate that our model is able to integrate noisy domain expertise in form of feature sets.
arXiv Detail & Related papers (2022-04-13T08:22:31Z) - TorchEsegeta: Framework for Interpretability and Explainability of
Image-based Deep Learning Models [0.0]
Clinicians are often sceptical about applying automatic image processing approaches, especially deep learning based methods, in practice.
This paper presents approaches that help to interpret and explain the results of deep learning algorithms by depicting the anatomical areas which influence the decision of the algorithm most.
Research presents a unified framework, TorchEsegeta, for applying various interpretability and explainability techniques for deep learning models.
arXiv Detail & Related papers (2021-10-16T01:00:15Z) - MAFER: a Multi-resolution Approach to Facial Expression Recognition [9.878384185493623]
We propose a two-step learning procedure, named MAFER, to train Deep Learning models tasked with recognizing facial expressions.
A relevant feature of MAFER is that it is task-agnostic, i.e., it can be used complementarily to other objective-related techniques.
arXiv Detail & Related papers (2021-05-06T07:26:58Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.