Single Cells Are Spatial Tokens: Transformers for Spatial Transcriptomic
Data Imputation
- URL: http://arxiv.org/abs/2302.03038v2
- Date: Fri, 16 Feb 2024 17:42:38 GMT
- Title: Single Cells Are Spatial Tokens: Transformers for Spatial Transcriptomic
Data Imputation
- Authors: Hongzhi Wen, Wenzhuo Tang, Wei Jin, Jiayuan Ding, Renming Liu, Xinnan
Dai, Feng Shi, Lulu Shang, Hui Liu, Yuying Xie
- Abstract summary: We present a transformer-based imputation framework, SpaFormer, for cellular-level spatial transcriptomic data.
We show that SpaFormer outperforms existing state-of-the-art imputation algorithms on three large-scale datasets.
- Score: 12.276966573615013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spatially resolved transcriptomics brings exciting breakthroughs to
single-cell analysis by providing physical locations along with gene
expression. However, as a cost of the extremely high spatial resolution, the
cellular level spatial transcriptomic data suffer significantly from missing
values. While a standard solution is to perform imputation on the missing
values, most existing methods either overlook spatial information or only
incorporate localized spatial context without the ability to capture long-range
spatial information. Using multi-head self-attention mechanisms and positional
encoding, transformer models can readily grasp the relationship between tokens
and encode location information. In this paper, by treating single cells as
spatial tokens, we study how to leverage transformers to facilitate spatial
tanscriptomics imputation. In particular, investigate the following two key
questions: (1) $\textit{how to encode spatial information of cells in
transformers}$, and (2) $\textit{ how to train a transformer for transcriptomic
imputation}$. By answering these two questions, we present a transformer-based
imputation framework, SpaFormer, for cellular-level spatial transcriptomic
data. Extensive experiments demonstrate that SpaFormer outperforms existing
state-of-the-art imputation algorithms on three large-scale datasets while
maintaining superior computational efficiency.
Related papers
- stEnTrans: Transformer-based deep learning for spatial transcriptomics enhancement [1.3124513975412255]
We present stEnTrans, a deep learning method based on Transformer architecture that provides comprehensive predictions for gene expression in unmeasured areas.
We evaluate stEnTrans on six datasets and the results indicate superior performance in enhancing spots resolution and predicting gene expression in unmeasured areas.
arXiv Detail & Related papers (2024-07-11T06:50:34Z) - stMCDI: Masked Conditional Diffusion Model with Graph Neural Network for Spatial Transcriptomics Data Imputation [8.211887623977214]
We introduce textbfstMCDI, a novel conditional diffusion model for spatial transcriptomics data imputation.
It employs a denoising network trained using randomly masked data portions as guidance, with the unmasked data serving as conditions.
The results obtained from spatial transcriptomics datasets elucidate the performance of our methods relative to existing approaches.
arXiv Detail & Related papers (2024-03-16T09:06:38Z) - Latent Space Inference For Spatial Transcriptomics [0.196629787330046]
We investigate a probabilistic machine learning method to obtain the full genetic expression information for tissues samples.
This is done through mapping both datasets to a joint latent space representation with the use of variational machine learning methods.
From here, the full genetic and spatial information can be decoded and to give us greater insights on the understanding of cellular processes and pathways.
arXiv Detail & Related papers (2023-11-01T06:50:00Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - Sub-token ViT Embedding via Stochastic Resonance Transformers [51.12001699637727]
Vision Transformer (ViT) architectures represent images as collections of high-dimensional vectorized tokens, each corresponding to a rectangular non-overlapping patch.
We propose a training-free method inspired by "stochastic resonance"
The resulting "Stochastic Resonance Transformer" (SRT) retains the rich semantic information of the original representation, but grounds it on a finer-scale spatial domain, partly mitigating the coarse effect of spatial tokenization.
arXiv Detail & Related papers (2023-10-06T01:53:27Z) - FactoFormer: Factorized Hyperspectral Transformers with Self-Supervised
Pretraining [36.44039681893334]
Hyperspectral images (HSIs) contain rich spectral and spatial information.
Current state-of-the-art hyperspectral transformers only tokenize the input HSI sample along the spectral dimension.
We propose a novel factorized spectral-spatial transformer that incorporates factorized self-supervised pretraining procedures.
arXiv Detail & Related papers (2023-09-18T02:05:52Z) - Focused Decoding Enables 3D Anatomical Detection by Transformers [64.36530874341666]
We propose a novel Detection Transformer for 3D anatomical structure detection, dubbed Focused Decoder.
Focused Decoder leverages information from an anatomical region atlas to simultaneously deploy query anchors and restrict the cross-attention's field of view.
We evaluate our proposed approach on two publicly available CT datasets and demonstrate that Focused Decoder not only provides strong detection results and thus alleviates the need for a vast amount of annotated data but also exhibits exceptional and highly intuitive explainability of results via attention weights.
arXiv Detail & Related papers (2022-07-21T22:17:21Z) - Towards Data-Efficient Detection Transformers [77.43470797296906]
We show most detection transformers suffer from significant performance drops on small-size datasets.
We empirically analyze the factors that affect data efficiency, through a step-by-step transition from a data-efficient RCNN variant to the representative DETR.
We introduce a simple yet effective label augmentation method to provide richer supervision and improve data efficiency.
arXiv Detail & Related papers (2022-03-17T17:56:34Z) - TraSeTR: Track-to-Segment Transformer with Contrastive Query for
Instance-level Instrument Segmentation in Robotic Surgery [60.439434751619736]
We propose TraSeTR, a Track-to-Segment Transformer that exploits tracking cues to assist surgical instrument segmentation.
TraSeTR jointly reasons about the instrument type, location, and identity with instance-level predictions.
The effectiveness of our method is demonstrated with state-of-the-art instrument type segmentation results on three public datasets.
arXiv Detail & Related papers (2022-02-17T05:52:18Z) - A Volumetric Transformer for Accurate 3D Tumor Segmentation [25.961484035609672]
This paper presents a Transformer architecture for medical image segmentation.
The Transformer has a U-shaped volumetric encoder-decoder design that processes the input voxels in their entirety.
We show that our model transfer better representations across-datasets and are robust against data corruptions.
arXiv Detail & Related papers (2021-11-26T02:49:51Z) - Transformers Solve the Limited Receptive Field for Monocular Depth
Prediction [82.90445525977904]
We propose TransDepth, an architecture which benefits from both convolutional neural networks and transformers.
This is the first paper which applies transformers into pixel-wise prediction problems involving continuous labels.
arXiv Detail & Related papers (2021-03-22T18:00:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.