Spatial encoding of BOLD fMRI time series for categorizing static images
across visual datasets: A pilot study on human vision
- URL: http://arxiv.org/abs/2309.03590v1
- Date: Thu, 7 Sep 2023 09:31:27 GMT
- Title: Spatial encoding of BOLD fMRI time series for categorizing static images
across visual datasets: A pilot study on human vision
- Authors: Vamshi K. Kancharala, Debanjali Bhattacharya and Neelam Sinha
- Abstract summary: Specific image categorization is performed using fMRI time series (TS) to understand differences in neuronal activities related to vision.
To understand vision, it is important to study how brain functions while looking at different images.
It is seen that parallel CNN model outperforms other network models with an improvement of 7% for multi-class classification.
- Score: 3.038642416291856
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Functional MRI (fMRI) is widely used to examine brain functionality by
detecting alteration in oxygenated blood flow that arises with brain activity.
In this study, complexity specific image categorization across different visual
datasets is performed using fMRI time series (TS) to understand differences in
neuronal activities related to vision. Publicly available BOLD5000 dataset is
used for this purpose, containing fMRI scans while viewing 5254 images of
diverse categories, drawn from three standard computer vision datasets: COCO,
ImageNet and SUN. To understand vision, it is important to study how brain
functions while looking at different images. To achieve this, spatial encoding
of fMRI BOLD TS has been performed that uses classical Gramian Angular Field
(GAF) and Markov Transition Field (MTF) to obtain 2D BOLD TS, representing
images of COCO, Imagenet and SUN. For classification, individual GAF and MTF
features are fed into regular CNN. Subsequently, parallel CNN model is employed
that uses combined 2D features for classifying images across COCO, Imagenet and
SUN. The result of 2D CNN models is also compared with 1D LSTM and Bi-LSTM that
utilizes raw fMRI BOLD signal for classification. It is seen that parallel CNN
model outperforms other network models with an improvement of 7% for
multi-class classification. Clinical relevance- The obtained result of this
analysis establishes a baseline in studying how differently human brain
functions while looking at images of diverse complexities.
Related papers
- MindFormer: Semantic Alignment of Multi-Subject fMRI for Brain Decoding [50.55024115943266]
We introduce a novel semantic alignment method of multi-subject fMRI signals using so-called MindFormer.
This model is specifically designed to generate fMRI-conditioned feature vectors that can be used for conditioning Stable Diffusion model for fMRI- to-image generation or large language model (LLM) for fMRI-to-text generation.
Our experimental results demonstrate that MindFormer generates semantically consistent images and text across different subjects.
arXiv Detail & Related papers (2024-05-28T00:36:25Z) - NeuroPictor: Refining fMRI-to-Image Reconstruction via Multi-individual Pretraining and Multi-level Modulation [55.51412454263856]
This paper proposes to directly modulate the generation process of diffusion models using fMRI signals.
By training with about 67,000 fMRI-image pairs from various individuals, our model enjoys superior fMRI-to-image decoding capacity.
arXiv Detail & Related papers (2024-03-27T02:42:52Z) - nnMamba: 3D Biomedical Image Segmentation, Classification and Landmark
Detection with State Space Model [24.955052600683423]
In this paper, we introduce nnMamba, a novel architecture that integrates the strengths of CNNs and the advanced long-range modeling capabilities of State Space Sequence Models (SSMs)
Experiments on 6 datasets demonstrate nnMamba's superiority over state-of-the-art methods in a suite of challenging tasks, including 3D image segmentation, classification, and landmark detection.
arXiv Detail & Related papers (2024-02-05T21:28:47Z) - Learning Multimodal Volumetric Features for Large-Scale Neuron Tracing [72.45257414889478]
We aim to reduce human workload by predicting connectivity between over-segmented neuron pieces.
We first construct a dataset, named FlyTracing, that contains millions of pairwise connections of segments expanding the whole fly brain.
We propose a novel connectivity-aware contrastive learning method to generate dense volumetric EM image embedding.
arXiv Detail & Related papers (2024-01-05T19:45:12Z) - Image complexity based fMRI-BOLD visual network categorization across
visual datasets using topological descriptors and deep-hybrid learning [3.522950356329991]
The aim of this study is to examine how network topology differs in response to distinct visual stimuli from visual datasets.
To achieve this, 0- and 1-dimensional persistence diagrams are computed for each visual network representing COCO, ImageNet, and SUN.
The extracted K-means cluster features are fed to a novel deep-hybrid model that yields accuracy in the range of 90%-95% in classifying these visual networks.
arXiv Detail & Related papers (2023-11-03T14:05:57Z) - Investigating the changes in BOLD responses during viewing of images
with varied complexity: An fMRI time-series based analysis on human vision [2.7036595757881323]
This work aims to investigate the neurological variation of human brain responses during viewing of images with varied complexity using fMRI time series (TS) analysis.
Our first study employs classical machine learning and deep learning strategies to classify image complexity-specific fMRI TS.
The obtained result of this analysis has established a baseline in studying how differently human brain functions while looking into images of diverse complexities.
arXiv Detail & Related papers (2023-09-27T08:46:09Z) - CoNeS: Conditional neural fields with shift modulation for multi-sequence MRI translation [5.662694302758443]
Multi-sequence magnetic resonance imaging (MRI) has found wide applications in both modern clinical studies and deep learning research.
It frequently occurs that one or more of the MRI sequences are missing due to different image acquisition protocols or contrast agent contraindications of patients.
One promising approach is to leverage generative models to synthesize the missing sequences, which can serve as a surrogate acquisition.
arXiv Detail & Related papers (2023-09-06T19:01:58Z) - JGAT: a joint spatio-temporal graph attention model for brain decoding [8.844033583141039]
Joint kernel Graph Attention Network (JGAT) is a new multi-modal temporal graph attention network framework.
It integrates the data from functional Magnetic Resonance Images (fMRI) and Diffusion Weighted Imaging (DWI) while preserving the dynamic information.
We conduct brain-decoding tasks with our JGAT on four independent datasets.
arXiv Detail & Related papers (2023-06-03T02:45:03Z) - BrainCLIP: Bridging Brain and Visual-Linguistic Representation Via CLIP
for Generic Natural Visual Stimulus Decoding [51.911473457195555]
BrainCLIP is a task-agnostic fMRI-based brain decoding model.
It bridges the modality gap between brain activity, image, and text.
BrainCLIP can reconstruct visual stimuli with high semantic fidelity.
arXiv Detail & Related papers (2023-02-25T03:28:54Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Comparisons among different stochastic selection of activation layers
for convolutional neural networks for healthcare [77.99636165307996]
We classify biomedical images using ensembles of neural networks.
We select our activations among the following ones: ReLU, leaky ReLU, Parametric ReLU, ELU, Adaptive Piecewice Linear Unit, S-Shaped ReLU, Swish, Mish, Mexican Linear Unit, Parametric Deformable Linear Unit, Soft Root Sign.
arXiv Detail & Related papers (2020-11-24T01:53:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.