Multi network InfoMax: A pre-training method involving graph
convolutional networks
- URL: http://arxiv.org/abs/2111.01276v1
- Date: Mon, 1 Nov 2021 21:53:20 GMT
- Title: Multi network InfoMax: A pre-training method involving graph
convolutional networks
- Authors: Usman Mahmood, Zening Fu, Vince Calhoun, Sergey Plis
- Abstract summary: This paper presents a pre-training method involving graph convolutional/neural networks (GCNs/GNNs)
The learned high-level graph latent representations help increase performance for downstream graph classification tasks.
We apply our method to a neuroimaging dataset for classifying subjects into healthy control (HC) and schizophrenia (SZ) groups.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Discovering distinct features and their relations from data can help us
uncover valuable knowledge crucial for various tasks, e.g., classification. In
neuroimaging, these features could help to understand, classify, and possibly
prevent brain disorders. Model introspection of highly performant
overparameterized deep learning (DL) models could help find these features and
relations. However, to achieve high-performance level DL models require
numerous labeled training samples ($n$) rarely available in many fields. This
paper presents a pre-training method involving graph convolutional/neural
networks (GCNs/GNNs), based on maximizing mutual information between two
high-level embeddings of an input sample. Many of the recently proposed
pre-training methods pre-train one of many possible networks of an
architecture. Since almost every DL model is an ensemble of multiple networks,
we take our high-level embeddings from two different networks of a model --a
convolutional and a graph network--. The learned high-level graph latent
representations help increase performance for downstream graph classification
tasks and bypass the need for a high number of labeled data samples. We apply
our method to a neuroimaging dataset for classifying subjects into healthy
control (HC) and schizophrenia (SZ) groups. Our experiments show that the
pre-trained model significantly outperforms the non-pre-trained model and
requires $50\%$ less data for similar performance.
Related papers
- GraphFM: A Scalable Framework for Multi-Graph Pretraining [2.882104808886318]
We introduce a scalable multi-graph multi-task pretraining approach specifically tailored for node classification tasks across diverse graph datasets from different domains.
We demonstrate the efficacy of our approach by training a model on 152 different graph datasets comprising over 7.4 million nodes and 189 million edges.
Our results show that pretraining on a diverse array of real and synthetic graphs improves the model's adaptability and stability, while performing competitively with state-of-the-art specialist models.
arXiv Detail & Related papers (2024-07-16T16:51:43Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Label Deconvolution for Node Representation Learning on Large-scale
Attributed Graphs against Learning Bias [75.44877675117749]
We propose an efficient label regularization technique, namely Label Deconvolution (LD), to alleviate the learning bias by a novel and highly scalable approximation to the inverse mapping of GNNs.
Experiments demonstrate LD significantly outperforms state-of-the-art methods on Open Graph datasets Benchmark.
arXiv Detail & Related papers (2023-09-26T13:09:43Z) - Layer-wise Linear Mode Connectivity [52.6945036534469]
Averaging neural network parameters is an intuitive method for the knowledge of two independent models.
It is most prominently used in federated learning.
We analyse the performance of the models that result from averaging single, or groups.
arXiv Detail & Related papers (2023-07-13T09:39:10Z) - Multi-Head Graph Convolutional Network for Structural Connectome
Classification [8.658134276685404]
We propose a machine-learning model inspired by graph convolutional networks (GCNs)
The proposed network is a simple design that employs different heads involving graph convolutions focused on edges and nodes.
To test the ability of our model to extract complementary and representative features from brain connectivity data, we chose the task of sex classification.
arXiv Detail & Related papers (2023-05-02T15:04:30Z) - Data-Free Adversarial Knowledge Distillation for Graph Neural Networks [62.71646916191515]
We propose the first end-to-end framework for data-free adversarial knowledge distillation on graph structured data (DFAD-GNN)
To be specific, our DFAD-GNN employs a generative adversarial network, which mainly consists of three components: a pre-trained teacher model and a student model are regarded as two discriminators, and a generator is utilized for deriving training graphs to distill knowledge from the teacher model into the student model.
Our DFAD-GNN significantly surpasses state-of-the-art data-free baselines in the graph classification task.
arXiv Detail & Related papers (2022-05-08T08:19:40Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - One Representative-Shot Learning Using a Population-Driven Template with
Application to Brain Connectivity Classification and Evolution Prediction [0.0]
Graph neural networks (GNNs) have been introduced to the field of network neuroscience.
We take a very different approach in training GNNs, where we aim to learn with one sample and achieve the best performance.
We present the first one-shot paradigm where a GNN is trained on a single population-driven template.
arXiv Detail & Related papers (2021-10-06T08:36:00Z) - Graph-Based Neural Network Models with Multiple Self-Supervised
Auxiliary Tasks [79.28094304325116]
Graph Convolutional Networks are among the most promising approaches for capturing relationships among structured data points.
We propose three novel self-supervised auxiliary tasks to train graph-based neural network models in a multi-task fashion.
arXiv Detail & Related papers (2020-11-14T11:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.