Sequential Transfer Learning to Decode Heard and Imagined Timbre from
fMRI Data
- URL: http://arxiv.org/abs/2305.13226v1
- Date: Mon, 22 May 2023 16:58:26 GMT
- Title: Sequential Transfer Learning to Decode Heard and Imagined Timbre from
fMRI Data
- Authors: Sean Paulsen, Michael Casey
- Abstract summary: We present a sequential transfer learning framework for transformers on functional Magnetic Resonance Imaging (fMRI) data.
In the first phase, we pre-train our stacked-encoder transformer architecture on Next Thought Prediction.
In the second phase, we fine-tune the models and train additional fresh models on the supervised task of predicting whether or not two sequences of fMRI data were recorded while listening to the same musical timbre.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a sequential transfer learning framework for transformers on
functional Magnetic Resonance Imaging (fMRI) data and demonstrate its
significant benefits for decoding musical timbre. In the first of two phases,
we pre-train our stacked-encoder transformer architecture on Next Thought
Prediction, a self-supervised task of predicting whether or not one sequence of
fMRI data follows another. This phase imparts a general understanding of the
temporal and spatial dynamics of neural activity, and can be applied to any
fMRI dataset. In the second phase, we fine-tune the pre-trained models and
train additional fresh models on the supervised task of predicting whether or
not two sequences of fMRI data were recorded while listening to the same
musical timbre. The fine-tuned models achieve significantly higher accuracy
with shorter training times than the fresh models, demonstrating the efficacy
of our framework for facilitating transfer learning on fMRI data. Additionally,
our fine-tuning task achieves a level of classification granularity beyond
standard methods. This work contributes to the growing literature on
transformer architectures for sequential transfer learning on fMRI data, and
provides evidence that our framework is an improvement over current methods for
decoding timbre.
Related papers
- Self-Supervised Pre-training Tasks for an fMRI Time-series Transformer in Autism Detection [3.665816629105171]
Autism Spectrum Disorder (ASD) is a neurodevelopmental condition that encompasses a wide variety of symptoms and degrees of impairment.
We have developed a transformer-based self-supervised framework that directly analyzes time-series fMRI data without computing functional connectivity.
We show that randomly masking entire ROIs gives better model performance than randomly masking time points in the pre-training step.
arXiv Detail & Related papers (2024-09-18T20:29:23Z) - Prompt Your Brain: Scaffold Prompt Tuning for Efficient Adaptation of fMRI Pre-trained Model [15.330413605539542]
Scaffold Prompt Tuning (ScaPT) is a novel prompt-based framework for adapting large-scale functional magnetic resonance imaging (fMRI) pre-trained models to downstream tasks.
It has high parameter efficiency and improved performance compared to fine-tuning and baselines for prompt tuning.
ScaPT outperforms fine-tuning and multitask-based prompt tuning in neurodegenerative diseases diagnosis/prognosis and personality trait prediction.
arXiv Detail & Related papers (2024-08-20T06:08:37Z) - NeuroPictor: Refining fMRI-to-Image Reconstruction via Multi-individual Pretraining and Multi-level Modulation [55.51412454263856]
This paper proposes to directly modulate the generation process of diffusion models using fMRI signals.
By training with about 67,000 fMRI-image pairs from various individuals, our model enjoys superior fMRI-to-image decoding capacity.
arXiv Detail & Related papers (2024-03-27T02:42:52Z) - Masked Audio Generation using a Single Non-Autoregressive Transformer [90.11646612273965]
MAGNeT is a masked generative sequence modeling method that operates directly over several streams of audio tokens.
We demonstrate the efficiency of MAGNeT for the task of text-to-music and text-to-audio generation.
We shed light on the importance of each of the components comprising MAGNeT, together with pointing to the trade-offs between autoregressive and non-autoregressive modeling.
arXiv Detail & Related papers (2024-01-09T14:29:39Z) - Skeleton2vec: A Self-supervised Learning Framework with Contextualized
Target Representations for Skeleton Sequence [56.092059713922744]
We show that using high-level contextualized features as prediction targets can achieve superior performance.
Specifically, we propose Skeleton2vec, a simple and efficient self-supervised 3D action representation learning framework.
Our proposed Skeleton2vec outperforms previous methods and achieves state-of-the-art results.
arXiv Detail & Related papers (2024-01-01T12:08:35Z) - Self-Supervised Pretraining on Paired Sequences of fMRI Data for
Transfer Learning to Brain Decoding Tasks [0.0]
We introduce a self-supervised pretraining framework for transformers on functional Magnetic Resonance Imaging (fMRI) data.
First, we pretrain our architecture on two self-supervised tasks simultaneously to teach the model a general understanding of the temporal and spatial dynamics of human auditory cortex during music listening.
Second, we finetune the pretrained models and train additional fresh models on a supervised fMRI classification task.
arXiv Detail & Related papers (2023-05-15T22:53:12Z) - Data-driven modelling of brain activity using neural networks, Diffusion
Maps, and the Koopman operator [0.0]
We propose a machine-learning approach to model long-term out-of-sample dynamics of brain activity from task-dependent fMRI data.
We use Diffusion maps (DMs) to discover a set of variables that parametrize the low-dimensional manifold on which the emergent high-dimensional fMRI time series evolve.
We construct reduced-order-models (ROMs) on the embedded manifold via two techniques: Feedforward Neural Networks (FNNs) and the Koopman operator.
arXiv Detail & Related papers (2023-04-24T09:08:12Z) - Pre-training and Fine-tuning Transformers for fMRI Prediction Tasks [69.85819388753579]
TFF employs a transformer-based architecture and a two-phase training approach.
Self-supervised training is applied to a collection of fMRI scans, where the model is trained for the reconstruction of 3D volume data.
Results show state-of-the-art performance on a variety of fMRI tasks, including age and gender prediction, as well as schizophrenia recognition.
arXiv Detail & Related papers (2021-12-10T18:04:26Z) - Reference-based Magnetic Resonance Image Reconstruction Using Texture
Transforme [86.6394254676369]
We propose a novel Texture Transformer Module (TTM) for accelerated MRI reconstruction.
We formulate the under-sampled data and reference data as queries and keys in a transformer.
The proposed TTM can be stacked on prior MRI reconstruction approaches to further improve their performance.
arXiv Detail & Related papers (2021-11-18T03:06:25Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.