MultiFIX: An XAI-friendly feature inducing approach to building models
from multimodal data
- URL: http://arxiv.org/abs/2402.12183v1
- Date: Mon, 19 Feb 2024 14:45:46 GMT
- Title: MultiFIX: An XAI-friendly feature inducing approach to building models
from multimodal data
- Authors: Mafalda Malafaia, Thalea Schlender, Peter A. N. Bosman, Tanja
Alderliesten
- Abstract summary: MultiFIX is a new interpretability-focused multimodal data fusion pipeline.
An end-to-end deep learning architecture is used to train a predictive model.
We apply MultiFIX to a publicly available dataset for the detection of malignant skin lesions.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the health domain, decisions are often based on different data modalities.
Thus, when creating prediction models, multimodal fusion approaches that can
extract and combine relevant features from different data modalities, can be
highly beneficial. Furthermore, it is important to understand how each modality
impacts the final prediction, especially in high-stake domains, so that these
models can be used in a trustworthy and responsible manner. We propose
MultiFIX: a new interpretability-focused multimodal data fusion pipeline that
explicitly induces separate features from different data types that can
subsequently be combined to make a final prediction. An end-to-end deep
learning architecture is used to train a predictive model and extract
representative features of each modality. Each part of the model is then
explained using explainable artificial intelligence techniques. Attention maps
are used to highlight important regions in image inputs. Inherently
interpretable symbolic expressions, learned with GP-GOMEA, are used to describe
the contribution of tabular inputs. The fusion of the extracted features to
predict the target label is also replaced by a symbolic expression, learned
with GP-GOMEA. Results on synthetic problems demonstrate the strengths and
limitations of MultiFIX. Lastly, we apply MultiFIX to a publicly available
dataset for the detection of malignant skin lesions.
Related papers
- TabDiff: a Multi-Modal Diffusion Model for Tabular Data Generation [91.50296404732902]
We introduce TabDiff, a joint diffusion framework that models all multi-modal distributions of tabular data in one model.
Our key innovation is the development of a joint continuous-time diffusion process for numerical and categorical data.
TabDiff achieves superior average performance over existing competitive baselines, with up to $22.5%$ improvement over the state-of-the-art model on pair-wise column correlation estimations.
arXiv Detail & Related papers (2024-10-27T22:58:47Z) - Flexible inference in heterogeneous and attributed multilayer networks [21.349513661012498]
We develop a probabilistic generative model to perform inference in multilayer networks with arbitrary types of information.
We demonstrate its ability to unveil a variety of patterns in a social support network among villagers in rural India.
arXiv Detail & Related papers (2024-05-31T15:21:59Z) - Debiasing Multimodal Models via Causal Information Minimization [65.23982806840182]
We study bias arising from confounders in a causal graph for multimodal data.
Robust predictive features contain diverse information that helps a model generalize to out-of-distribution data.
We use these features as confounder representations and use them via methods motivated by causal theory to remove bias from models.
arXiv Detail & Related papers (2023-11-28T16:46:14Z) - HEALNet: Multimodal Fusion for Heterogeneous Biomedical Data [10.774128925670183]
This paper presents the Hybrid Early-fusion Attention Learning Network (HEALNet), a flexible multimodal fusion architecture.
We conduct multimodal survival analysis on Whole Slide Images and Multi-omic data on four cancer datasets from The Cancer Genome Atlas (TCGA)
HEALNet achieves state-of-the-art performance compared to other end-to-end trained fusion models.
arXiv Detail & Related papers (2023-11-15T17:06:26Z) - End-to-End Learning on Multimodal Knowledge Graphs [0.0]
We propose a multimodal message passing network which learns end-to-end from the structure of graphs.
Our model uses dedicated (neural) encoders to naturally learn embeddings for node features belonging to five different types of modalities.
Our results indicate that end-to-end multimodal learning from any arbitrary knowledge graph is indeed possible.
arXiv Detail & Related papers (2023-09-03T13:16:18Z) - StableLLaVA: Enhanced Visual Instruction Tuning with Synthesized
Image-Dialogue Data [129.92449761766025]
We propose a novel data collection methodology that synchronously synthesizes images and dialogues for visual instruction tuning.
This approach harnesses the power of generative models, marrying the abilities of ChatGPT and text-to-image generative models.
Our research includes comprehensive experiments conducted on various datasets.
arXiv Detail & Related papers (2023-08-20T12:43:52Z) - Dynamic Latent Separation for Deep Learning [67.62190501599176]
A core problem in machine learning is to learn expressive latent variables for model prediction on complex data.
Here, we develop an approach that improves expressiveness, provides partial interpretation, and is not restricted to specific applications.
arXiv Detail & Related papers (2022-10-07T17:56:53Z) - Encoding Domain Knowledge in Multi-view Latent Variable Models: A
Bayesian Approach with Structured Sparsity [7.811916700683125]
MuVI is a novel approach for domain-informed multi-view latent variable models.
We demonstrate that our model is able to integrate noisy domain expertise in form of feature sets.
arXiv Detail & Related papers (2022-04-13T08:22:31Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Connecting the Dots: Multivariate Time Series Forecasting with Graph
Neural Networks [91.65637773358347]
We propose a general graph neural network framework designed specifically for multivariate time series data.
Our approach automatically extracts the uni-directed relations among variables through a graph learning module.
Our proposed model outperforms the state-of-the-art baseline methods on 3 of 4 benchmark datasets.
arXiv Detail & Related papers (2020-05-24T04:02:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.