CellViT: Vision Transformers for Precise Cell Segmentation and
Classification
- URL: http://arxiv.org/abs/2306.15350v2
- Date: Fri, 6 Oct 2023 13:22:14 GMT
- Title: CellViT: Vision Transformers for Precise Cell Segmentation and
Classification
- Authors: Fabian H\"orst, Moritz Rempe, Lukas Heine, Constantin Seibold, Julius
Keyl, Giulia Baldini, Selma Ugurel, Jens Siveke, Barbara Gr\"unwald, Jan
Egger, Jens Kleesiek
- Abstract summary: We introduce a new method for automated instance segmentation of cell nuclei in digitized tissue samples using a deep learning architecture based on Vision Transformer called CellViT.
We demonstrate the superiority of large-scale in-domain and out-of-domain pre-trained Vision Transformers by leveraging the recently published Segment Anything Model and a ViT-encoder pre-trained on 104 million histological image patches.
- Score: 3.6000652088960785
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Nuclei detection and segmentation in hematoxylin and eosin-stained (H&E)
tissue images are important clinical tasks and crucial for a wide range of
applications. However, it is a challenging task due to nuclei variances in
staining and size, overlapping boundaries, and nuclei clustering. While
convolutional neural networks have been extensively used for this task, we
explore the potential of Transformer-based networks in this domain. Therefore,
we introduce a new method for automated instance segmentation of cell nuclei in
digitized tissue samples using a deep learning architecture based on Vision
Transformer called CellViT. CellViT is trained and evaluated on the PanNuke
dataset, which is one of the most challenging nuclei instance segmentation
datasets, consisting of nearly 200,000 annotated Nuclei into 5 clinically
important classes in 19 tissue types. We demonstrate the superiority of
large-scale in-domain and out-of-domain pre-trained Vision Transformers by
leveraging the recently published Segment Anything Model and a ViT-encoder
pre-trained on 104 million histological image patches - achieving
state-of-the-art nuclei detection and instance segmentation performance on the
PanNuke dataset with a mean panoptic quality of 0.50 and an F1-detection score
of 0.83. The code is publicly available at https://github.com/TIO-IKIM/CellViT
Related papers
- Cell Nuclei Detection and Classification in Whole Slide Images with Transformers [1.7129141499083573]
We show that CellNuc-DETR is twice as fast as the fastest segmentation-based method, HoVer-NeXt, while achieving substantially higher accuracy.
It surpasses CellViT in accuracy and is approximately ten times more efficient in inference speed on WSIs.
arXiv Detail & Related papers (2025-02-10T09:52:02Z) - Channel Boosted CNN-Transformer-based Multi-Level and Multi-Scale Nuclei Segmentation [0.40964539027092917]
nuclei segmentation is an essential foundation for various applications in computational pathology, including cancer diagnosis and treatment planning.
achieving accurate segmentation remains challenging due to factors like clustered nuclei, high intra-class variability in size and shape, resemblance to other cells, and color or contrast variations between nuclei and background.
We propose two CNN-Transformer architectures that leverage the strengths of both CNNs and Transformers to effectively learn nuclei boundaries in multi-organ histology images.
arXiv Detail & Related papers (2024-07-27T05:54:05Z) - LKCell: Efficient Cell Nuclei Instance Segmentation with Large Convolution Kernels [32.157968641130545]
We propose LKCell, a high-accuracy and efficient cell segmentation method.
Its core insight lies in unleashing the potential of large convolution kernels to achieve computationally efficient large receptive fields.
We analyze the redundancy of previous methods and design a new segmentation decoder based on large convolution kernels.
arXiv Detail & Related papers (2024-07-25T14:07:49Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Domain Adaptive Nuclei Instance Segmentation and Classification via
Category-aware Feature Alignment and Pseudo-labelling [65.40672505658213]
We propose a novel deep neural network, namely Category-Aware feature alignment and Pseudo-Labelling Network (CAPL-Net) for UDA nuclei instance segmentation and classification.
Our approach outperforms state-of-the-art UDA methods with a remarkable margin.
arXiv Detail & Related papers (2022-07-04T07:05:06Z) - Global Context Vision Transformers [78.5346173956383]
We propose global context vision transformer (GC ViT), a novel architecture that enhances parameter and compute utilization for computer vision.
We address the lack of the inductive bias in ViTs, and propose to leverage a modified fused inverted residual blocks in our architecture.
Our proposed GC ViT achieves state-of-the-art results across image classification, object detection and semantic segmentation tasks.
arXiv Detail & Related papers (2022-06-20T18:42:44Z) - PointNu-Net: Keypoint-assisted Convolutional Neural Network for
Simultaneous Multi-tissue Histology Nuclei Segmentation and Classification [23.466331358975044]
We study and design a new method to simultaneously detect, segment, and classify nuclei from Haematoxylin and Eosin stained histopathology data.
We demonstrate the superior performance of our proposed approach for nuclei segmentation and classification across 19 different tissue types.
arXiv Detail & Related papers (2021-11-01T08:29:40Z) - Towards an Automatic Analysis of CHO-K1 Suspension Growth in
Microfluidic Single-cell Cultivation [63.94623495501023]
We propose a novel Machine Learning architecture, which allows us to infuse a neural deep network with human-powered abstraction on the level of data.
Specifically, we train a generative model simultaneously on natural and synthetic data, so that it learns a shared representation, from which a target variable, such as the cell count, can be reliably estimated.
arXiv Detail & Related papers (2020-10-20T08:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.