Few-shot Medical Image Segmentation via Cross-Reference Transformer
- URL: http://arxiv.org/abs/2304.09630v4
- Date: Wed, 26 Jul 2023 11:55:03 GMT
- Title: Few-shot Medical Image Segmentation via Cross-Reference Transformer
- Authors: Yao Huang and Jianming Liu
- Abstract summary: Few-shot segmentation(FSS) has the potential to address these challenges by learning new categories from a small number of labeled samples.
We propose a novel self-supervised few shot medical image segmentation network with Cross-Reference Transformer.
Experimental results show that the proposed model achieves good results on both CT dataset and MRI dataset.
- Score: 3.2634122554914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning models have become the mainstream method for medical image
segmentation, but they require a large manually labeled dataset for training
and are difficult to extend to unseen categories. Few-shot segmentation(FSS)
has the potential to address these challenges by learning new categories from a
small number of labeled samples. The majority of the current methods employ a
prototype learning architecture, which involves expanding support prototype
vectors and concatenating them with query features to conduct conditional
segmentation. However, such framework potentially focuses more on query
features while may neglect the correlation between support and query features.
In this paper, we propose a novel self-supervised few shot medical image
segmentation network with Cross-Reference Transformer, which addresses the lack
of interaction between the support image and the query image. We first enhance
the correlation features between the support set image and the query image
using a bidirectional cross-attention module. Then, we employ a cross-reference
mechanism to mine and enhance the similar parts of support features and query
features in high-dimensional channels. Experimental results show that the
proposed model achieves good results on both CT dataset and MRI dataset.
Related papers
- Correlation Weighted Prototype-based Self-Supervised One-Shot Segmentation of Medical Images [12.365801596593936]
Medical image segmentation is one of the domains where sufficient annotated data is not available.
We propose a prototype-based self-supervised one-way one-shot learning framework using pseudo-labels generated from superpixels.
We show that the proposed simple but potent framework performs at par with the state-of-the-art methods.
arXiv Detail & Related papers (2024-08-12T15:38:51Z) - Cross-Domain Few-Shot Semantic Segmentation via Doubly Matching Transformation [26.788260801305974]
Cross-Domain Few-shot Semantic (CD-FSS) aims to train generalized models that can segment classes from different domains with a few labeled images.
Previous works have proven the effectiveness of feature transformation in addressing CD-FSS.
We propose a Doubly Matching Transformation-based Network (DMTNet) to solve the above issue.
arXiv Detail & Related papers (2024-05-24T06:47:43Z) - Support-Query Prototype Fusion Network for Few-shot Medical Image Segmentation [7.6695642174485705]
Few-shot learning, which utilizes a small amount of labeled data to generalize to unseen classes, has emerged as a critical research area.
We propose a novel Support-Query Prototype Fusion Network (SQPFNet) to mitigate this drawback.
evaluation results on two public datasets, SABS and CMR, show that SQPFNet achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-05-13T07:31:16Z) - MatchSeg: Towards Better Segmentation via Reference Image Matching [5.55078598520531]
Few-shot learning aims to overcome the need for annotated data by using a small labeled dataset, known as a support set, to guide predicting labels for new, unlabeled images.
Inspired by this paradigm, we introduce MatchSeg, a novel framework that enhances medical image segmentation through strategic reference image matching.
arXiv Detail & Related papers (2024-03-23T18:04:58Z) - Multi-grained Temporal Prototype Learning for Few-shot Video Object
Segmentation [156.4142424784322]
Few-Shot Video Object (FSVOS) aims to segment objects in a query video with the same category defined by a few annotated support images.
We propose to leverage multi-grained temporal guidance information for handling the temporal correlation nature of video data.
Our proposed video IPMT model significantly outperforms previous models on two benchmark datasets.
arXiv Detail & Related papers (2023-09-20T09:16:34Z) - Few Shot Medical Image Segmentation with Cross Attention Transformer [30.54965157877615]
We propose a novel framework for few-shot medical image segmentation, termed CAT-Net.
Our proposed network mines the correlations between the support image and query image, limiting them to focus only on useful foreground information.
We validated the proposed method on three public datasets: Abd-CT, Abd-MRI, and Card-MRI.
arXiv Detail & Related papers (2023-03-24T09:10:14Z) - Intermediate Prototype Mining Transformer for Few-Shot Semantic
Segmentation [119.51445225693382]
Few-shot semantic segmentation aims to segment the target objects in query under the condition of a few annotated support images.
We introduce an intermediate prototype for mining both deterministic category information from the support and adaptive category knowledge from the query.
In each IPMT layer, we propagate the object information in both support and query features to the prototype and then use it to activate the query feature map.
arXiv Detail & Related papers (2022-10-13T06:45:07Z) - Few-shot Segmentation with Optimal Transport Matching and Message Flow [50.9853556696858]
It is essential for few-shot semantic segmentation to fully utilize the support information.
We propose a Correspondence Matching Network (CMNet) with an Optimal Transport Matching module.
Experiments on PASCAL VOC 2012, MS COCO, and FSS-1000 datasets show that our network achieves new state-of-the-art few-shot segmentation performance.
arXiv Detail & Related papers (2021-08-19T06:26:11Z) - Few-Shot Segmentation via Cycle-Consistent Transformer [74.49307213431952]
We focus on utilizing pixel-wise relationships between support and target images to facilitate the few-shot semantic segmentation task.
We propose using a novel cycle-consistent attention mechanism to filter out possible harmful support features.
Our proposed CyCTR leads to remarkable improvement compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-06-04T07:57:48Z) - BriNet: Towards Bridging the Intra-class and Inter-class Gaps in
One-Shot Segmentation [84.2925550033094]
Few-shot segmentation focuses on the generalization of models to segment unseen object instances with limited training samples.
We propose a framework, BriNet, to bridge the gaps between the extracted features of the query and support images.
The effectiveness of our framework is demonstrated by experimental results, which outperforms other competitive methods.
arXiv Detail & Related papers (2020-08-14T07:45:50Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.