Multi-View Spatial-Temporal Graph Convolutional Networks with Domain
Generalization for Sleep Stage Classification
- URL: http://arxiv.org/abs/2109.01824v1
- Date: Sat, 4 Sep 2021 09:19:27 GMT
- Title: Multi-View Spatial-Temporal Graph Convolutional Networks with Domain
Generalization for Sleep Stage Classification
- Authors: Ziyu Jia, Youfang Lin, Jing Wang, Xiaojun Ning, Yuanlai He, Ronghao
Zhou, Yuhan Zhou, Li-wei H. Lehman
- Abstract summary: We propose a multi-view spatial-temporal graph convolutional networks (MSTGCN) with domain generalization for sleep stage classification.
MSTGCN consists of graph convolutions for extracting spatial features and temporal convolutions for capturing the transition rules among sleep stages.
Experiments on two public datasets demonstrate that the proposed model outperforms the state-of-the-art baselines.
- Score: 12.488891852531422
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sleep stage classification is essential for sleep assessment and disease
diagnosis. Although previous attempts to classify sleep stages have achieved
high classification performance, several challenges remain open: 1) How to
effectively utilize time-varying spatial and temporal features from
multi-channel brain signals remains challenging. Prior works have not been able
to fully utilize the spatial topological information among brain regions. 2)
Due to the many differences found in individual biological signals, how to
overcome the differences of subjects and improve the generalization of deep
neural networks is important. 3) Most deep learning methods ignore the
interpretability of the model to the brain. To address the above challenges, we
propose a multi-view spatial-temporal graph convolutional networks (MSTGCN)
with domain generalization for sleep stage classification. Specifically, we
construct two brain view graphs for MSTGCN based on the functional connectivity
and physical distance proximity of the brain regions. The MSTGCN consists of
graph convolutions for extracting spatial features and temporal convolutions
for capturing the transition rules among sleep stages. In addition, attention
mechanism is employed for capturing the most relevant spatial-temporal
information for sleep stage classification. Finally, domain generalization and
MSTGCN are integrated into a unified framework to extract subject-invariant
sleep features. Experiments on two public datasets demonstrate that the
proposed model outperforms the state-of-the-art baselines.
Related papers
- Predicting Infant Brain Connectivity with Federated Multi-Trajectory
GNNs using Scarce Data [54.55126643084341]
Existing deep learning solutions suffer from three major limitations.
We introduce FedGmTE-Net++, a federated graph-based multi-trajectory evolution network.
Using the power of federation, we aggregate local learnings among diverse hospitals with limited datasets.
arXiv Detail & Related papers (2024-01-01T10:20:01Z) - HDGL: A hierarchical dynamic graph representation learning model for
brain disorder classification [1.7495515703051119]
We propose a hierarchical dynamic graph representation learning (HDGL) model, which is the first model designed to address all the aforementioned challenges.
We evaluate the performance of the proposed model on the ABIDE and ADHD-200 datasets.
arXiv Detail & Related papers (2023-11-06T06:29:23Z) - DepWiGNN: A Depth-wise Graph Neural Network for Multi-hop Spatial
Reasoning in Text [52.699307699505646]
We propose a novel Depth-Wise Graph Neural Network (DepWiGNN) to handle multi-hop spatial reasoning.
Specifically, we design a novel node memory scheme and aggregate the information over the depth dimension instead of the breadth dimension of the graph.
Experimental results on two challenging multi-hop spatial reasoning datasets show that DepWiGNN outperforms existing spatial reasoning methods.
arXiv Detail & Related papers (2023-10-19T08:07:22Z) - Language Knowledge-Assisted Representation Learning for Skeleton-Based
Action Recognition [71.35205097460124]
How humans understand and recognize the actions of others is a complex neuroscientific problem.
LA-GCN proposes a graph convolution network using large-scale language models (LLM) knowledge assistance.
arXiv Detail & Related papers (2023-05-21T08:29:16Z) - DBGDGM: Dynamic Brain Graph Deep Generative Model [63.23390833353625]
Graphs are a natural representation of brain activity derived from functional magnetic imaging (fMRI) data.
It is well known that clusters of anatomical brain regions, known as functional connectivity networks (FCNs), encode temporal relationships which can serve as useful biomarkers for understanding brain function and dysfunction.
Previous works, however, ignore the temporal dynamics of the brain and focus on static graphs.
We propose a dynamic brain graph deep generative model (DBGDGM) which simultaneously clusters brain regions into temporally evolving communities and learns dynamic unsupervised node embeddings.
arXiv Detail & Related papers (2023-01-26T20:45:30Z) - ProductGraphSleepNet: Sleep Staging using Product Spatio-Temporal Graph
Learning with Attentive Temporal Aggregation [4.014524824655106]
This work proposes an adaptive product graph learning-based graph convolutional network, named ProductGraphSleepNet, for learning joint-temporal graphs.
The proposed network makes it possible for clinicians to comprehend and interpret the learned connectivity graphs for sleep stages.
arXiv Detail & Related papers (2022-12-09T14:34:58Z) - SleePyCo: Automatic Sleep Scoring with Feature Pyramid and Contrastive
Learning [0.0]
We propose a deep learning framework named SleePyCo that incorporates 1) a feature pyramid and 2) supervised contrastive learning for automatic sleep scoring.
For the feature pyramid, we propose a backbone network named SleePyCo-backbone to consider multiple feature sequences on different temporal and frequency scales.
Supervised contrastive learning allows the network to extract class discriminative features by minimizing the distance between intra-class features and simultaneously maximizing that between inter-class features.
arXiv Detail & Related papers (2022-09-20T04:10:49Z) - Self-Supervised Graph Representation Learning for Neuronal Morphologies [75.38832711445421]
We present GraphDINO, a data-driven approach to learn low-dimensional representations of 3D neuronal morphologies from unlabeled datasets.
We show, in two different species and across multiple brain areas, that this method yields morphological cell type clusterings on par with manual feature-based classification by experts.
Our method could potentially enable data-driven discovery of novel morphological features and cell types in large-scale datasets.
arXiv Detail & Related papers (2021-12-23T12:17:47Z) - Identifying Autism Spectrum Disorder Based on Individual-Aware
Down-Sampling and Multi-Modal Learning [4.310840361752551]
We propose a novel feature extraction method for fMRI that can learn a personalized lowe-resolution representation of the entire brain networking.
The present model has achieved a mean classification accuracy of 85.95% and a mean AUC of 0.92, which is better than the state-of-the-art methods.
arXiv Detail & Related papers (2021-09-19T14:22:55Z) - Spatial-Temporal Correlation and Topology Learning for Person
Re-Identification in Videos [78.45050529204701]
We propose a novel framework to pursue discriminative and robust representation by modeling cross-scale spatial-temporal correlation.
CTL utilizes a CNN backbone and a key-points estimator to extract semantic local features from human body.
It explores a context-reinforced topology to construct multi-scale graphs by considering both global contextual information and physical connections of human body.
arXiv Detail & Related papers (2021-04-15T14:32:12Z) - Convolutional Neural Networks for cytoarchitectonic brain mapping at
large scale [0.33727511459109777]
We present a new workflow for mapping cytoarchitectonic areas in large series of cell-body stained histological sections of human postmortem brains.
It is based on a Deep Convolutional Neural Network (CNN), which is trained on a pair of section images with annotations, with a large number of un-annotated sections in between.
The new workflow does not require preceding 3D-reconstruction of sections, and is robust against histological artefacts.
arXiv Detail & Related papers (2020-11-25T16:25:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.