A Few-shot Learning Graph Multi-Trajectory Evolution Network for
Forecasting Multimodal Baby Connectivity Development from a Baseline
Timepoint
- URL: http://arxiv.org/abs/2110.03535v1
- Date: Wed, 6 Oct 2021 08:26:57 GMT
- Title: A Few-shot Learning Graph Multi-Trajectory Evolution Network for
Forecasting Multimodal Baby Connectivity Development from a Baseline
Timepoint
- Authors: Alaa Bessadok, Ahmed Nebli, Mohamed Ali Mahjoub, Gang Li, Weili Lin,
Dinggang Shen and Islem Rekik
- Abstract summary: We propose a Graph Multi-Trajectory Evolution Network (GmTE-Net), which adopts a teacher-student paradigm.
This is the first teacher-student architecture tailored for brain graph multi-trajectory growth prediction.
- Score: 53.73316520733503
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Charting the baby connectome evolution trajectory during the first year after
birth plays a vital role in understanding dynamic connectivity development of
baby brains. Such analysis requires acquisition of longitudinal connectomic
datasets. However, both neonatal and postnatal scans are rarely acquired due to
various difficulties. A small body of works has focused on predicting baby
brain evolution trajectory from a neonatal brain connectome derived from a
single modality. Although promising, large training datasets are essential to
boost model learning and to generalize to a multi-trajectory prediction from
different modalities (i.e., functional and morphological connectomes). Here, we
unprecedentedly explore the question: Can we design a few-shot learning-based
framework for predicting brain graph trajectories across different modalities?
To this aim, we propose a Graph Multi-Trajectory Evolution Network (GmTE-Net),
which adopts a teacher-student paradigm where the teacher network learns on
pure neonatal brain graphs and the student network learns on simulated brain
graphs given a set of different timepoints. To the best of our knowledge, this
is the first teacher-student architecture tailored for brain graph
multi-trajectory growth prediction that is based on few-shot learning and
generalized to graph neural networks (GNNs). To boost the performance of the
student network, we introduce a local topology-aware distillation loss that
forces the predicted graph topology of the student network to be consistent
with the teacher network. Experimental results demonstrate substantial
performance gains over benchmark methods. Hence, our GmTE-Net can be leveraged
to predict atypical brain connectivity trajectory evolution across various
modalities. Our code is available at https: //github.com/basiralab/GmTE-Net.
Related papers
- Graph Neural Networks for Brain Graph Learning: A Survey [53.74244221027981]
Graph neural networks (GNNs) have demonstrated a significant advantage in mining graph-structured data.
GNNs to learn brain graph representations for brain disorder analysis has recently gained increasing attention.
In this paper, we aim to bridge this gap by reviewing brain graph learning works that utilize GNNs.
arXiv Detail & Related papers (2024-06-01T02:47:39Z) - Predicting Infant Brain Connectivity with Federated Multi-Trajectory
GNNs using Scarce Data [54.55126643084341]
Existing deep learning solutions suffer from three major limitations.
We introduce FedGmTE-Net++, a federated graph-based multi-trajectory evolution network.
Using the power of federation, we aggregate local learnings among diverse hospitals with limited datasets.
arXiv Detail & Related papers (2024-01-01T10:20:01Z) - Multi-Head Graph Convolutional Network for Structural Connectome
Classification [8.658134276685404]
We propose a machine-learning model inspired by graph convolutional networks (GCNs)
The proposed network is a simple design that employs different heads involving graph convolutions focused on edges and nodes.
To test the ability of our model to extract complementary and representative features from brain connectivity data, we chose the task of sex classification.
arXiv Detail & Related papers (2023-05-02T15:04:30Z) - Contrastive Brain Network Learning via Hierarchical Signed Graph Pooling
Model [64.29487107585665]
Graph representation learning techniques on brain functional networks can facilitate the discovery of novel biomarkers for clinical phenotypes and neurodegenerative diseases.
Here, we propose an interpretable hierarchical signed graph representation learning model to extract graph-level representations from brain functional networks.
In order to further improve the model performance, we also propose a new strategy to augment functional brain network data for contrastive learning.
arXiv Detail & Related papers (2022-07-14T20:03:52Z) - Inter-Domain Alignment for Predicting High-Resolution Brain Networks
Using Teacher-Student Learning [0.0]
We propose Learn to SuperResolve Brain Graphs with Knowledge Distillation Network (L2S-KDnet) to superresolve brain graphs.
Our teacher network is a graph encoder-decoder that firstly learns the LR brain graph embeddings, and secondly learns how to align the resulting latent representations to the HR ground truth data distribution.
Next, our student network learns the knowledge of the aligned brain graphs as well as the topological structure of the predicted HR graphs transferred from the teacher.
arXiv Detail & Related papers (2021-10-06T09:31:44Z) - Foreseeing Brain Graph Evolution Over Time Using Deep Adversarial
Network Normalizer [0.0]
We propose the first graph-based Generative Adversarial Network (gGAN) that learns how to normalize brain graphs.
Our proposed method achieved the lowest brain disease evolution prediction error using a single baseline timepoint.
arXiv Detail & Related papers (2020-09-23T14:25:40Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z) - The large learning rate phase of deep learning: the catapult mechanism [50.23041928811575]
We present a class of neural networks with solvable training dynamics.
We find good agreement between our model's predictions and training dynamics in realistic deep learning settings.
We believe our results shed light on characteristics of models trained at different learning rates.
arXiv Detail & Related papers (2020-03-04T17:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.