Multi-modal Graph Learning for Disease Prediction
- URL: http://arxiv.org/abs/2107.00206v1
- Date: Thu, 1 Jul 2021 03:59:22 GMT
- Title: Multi-modal Graph Learning for Disease Prediction
- Authors: Shuai Zheng, Zhenfeng Zhu, Zhizhe Liu, Zhenyu Guo, Yang Liu, Yao Zhao
- Abstract summary: We propose an end-to-end Multimodal Graph Learning framework (MMGL) for disease prediction.
Instead of defining the adjacency matrix manually as existing methods, the latent graph structure can be captured through a novel way of adaptive graph learning.
- Score: 35.4310911850558
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Benefiting from the powerful expressive capability of graphs, graph-based
approaches have achieved impressive performance in various biomedical
applications. Most existing methods tend to define the adjacency matrix among
samples manually based on meta-features, and then obtain the node embeddings
for downstream tasks by Graph Representation Learning (GRL). However, it is not
easy for these approaches to generalize to unseen samples. Meanwhile, the
complex correlation between modalities is also ignored. As a result, these
factors inevitably yield the inadequacy of providing valid information about
the patient's condition for a reliable diagnosis. In this paper, we propose an
end-to-end Multimodal Graph Learning framework (MMGL) for disease prediction.
To effectively exploit the rich information across multi-modality associated
with diseases, amodal-attentional multi-modal fusion is proposed to integrate
the features of each modality by leveraging the correlation and complementarity
between the modalities. Furthermore, instead of defining the adjacency matrix
manually as existing methods, the latent graph structure can be captured
through a novel way of adaptive graph learning. It could be jointly optimized
with the prediction model, thus revealing the intrinsic connections among
samples. Unlike the previous transductive methods, our model is also applicable
to the scenario of inductive learning for those unseen data. An extensive group
of experiments on two disease prediction problems is then carefully designed
and presented, demonstrating that MMGL obtains more favorable performances. In
addition, we also visualize and analyze the learned graph structure to provide
more reliable decision support for doctors in real medical applications and
inspiration for disease research.
Related papers
- GTP-4o: Modality-prompted Heterogeneous Graph Learning for Omni-modal Biomedical Representation [68.63955715643974]
Modality-prompted Heterogeneous Graph for Omnimodal Learning (GTP-4o)
We propose an innovative Modality-prompted Heterogeneous Graph for Omnimodal Learning (GTP-4o)
arXiv Detail & Related papers (2024-07-08T01:06:13Z) - MM-GTUNets: Unified Multi-Modal Graph Deep Learning for Brain Disorders Prediction [8.592259720470697]
We propose MM-GTUNets, an end-to-end graph transformer based multi-modal graph deep learning framework for brain disorders prediction.
We introduce Modality Reward Representation Learning (MRRL) which adaptively constructs population graphs using a reward system.
We also propose Adaptive Cross-Modal Graph Learning (ACMGL), which captures critical modality-specific and modality-shared features.
arXiv Detail & Related papers (2024-06-20T16:14:43Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - MaxCorrMGNN: A Multi-Graph Neural Network Framework for Generalized
Multimodal Fusion of Medical Data for Outcome Prediction [3.2889220522843625]
We develop an innovative fusion approach called MaxCorr MGNN that models non-linear modality correlations within and across patients.
We then design, for the first time, a generalized multi-layered graph neural network (MGNN) for task-informed reasoning in multi-layered graphs.
We evaluate our model an outcome prediction task on a Tuberculosis dataset consistently outperforming several state-of-the-art neural, graph-based and traditional fusion techniques.
arXiv Detail & Related papers (2023-07-13T23:52:41Z) - Multi-modal Multi-kernel Graph Learning for Autism Prediction and
Biomarker Discovery [29.790200009136825]
We propose a novel method to offset the negative impact between modalities in the process of multi-modal integration and extract heterogeneous information from graphs.
Our method is evaluated on the benchmark Autism Brain Imaging Data Exchange (ABIDE) dataset and outperforms the state-of-the-art methods.
In addition, discriminative brain regions associated with autism are identified by our model, providing guidance for the study of autism pathology.
arXiv Detail & Related papers (2023-03-03T07:09:17Z) - Fusing Modalities by Multiplexed Graph Neural Networks for Outcome
Prediction in Tuberculosis [3.131872070347212]
We present a novel fusion framework using multiplexed graphs and derive a new graph neural network for learning from such graphs.
We present results that show that our proposed method outperforms state-of-the-art methods of fusing modalities for multi-outcome prediction on a large Tuberculosis (TB) dataset.
arXiv Detail & Related papers (2022-10-25T23:03:05Z) - Multi-modal Graph Learning for Disease Prediction [35.156975779372836]
We propose an end-to-end Multi-modal Graph Learning framework (MMGL) for disease prediction with multi-modality.
Instead of defining the graph manually, the latent graph structure is captured through an effective way of adaptive graph learning.
An extensive group of experiments on two disease prediction tasks demonstrates that the proposed MMGL achieves more favorable performance.
arXiv Detail & Related papers (2022-03-11T12:33:20Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Dynamic Graph Correlation Learning for Disease Diagnosis with Incomplete
Labels [66.57101219176275]
Disease diagnosis on chest X-ray images is a challenging multi-label classification task.
We propose a Disease Diagnosis Graph Convolutional Network (DD-GCN) that presents a novel view of investigating the inter-dependency among different diseases.
Our method is the first to build a graph over the feature maps with a dynamic adjacency matrix for correlation learning.
arXiv Detail & Related papers (2020-02-26T17:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.