vMFNet: Compositionality Meets Domain-generalised Segmentation
- URL: http://arxiv.org/abs/2206.14538v1
- Date: Wed, 29 Jun 2022 11:31:23 GMT
- Title: vMFNet: Compositionality Meets Domain-generalised Segmentation
- Authors: Xiao Liu, Spyridon Thermos, Pedro Sanchez, Alison Q. O'Neil and
Sotirios A. Tsaftaris
- Abstract summary: von-Mises-Fisher (vMF) kernels are robust to images collected from different domains.
The vMF likelihoods tell how likely each anatomical part is at each position of the image.
With a reconstruction module, unlabeled data can also be used to learn the vMF kernels and likelihoods.
- Score: 22.741376970643973
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Training medical image segmentation models usually requires a large amount of
labeled data. By contrast, humans can quickly learn to accurately recognise
anatomy of interest from medical (e.g. MRI and CT) images with some limited
guidance. Such recognition ability can easily generalise to new images from
different clinical centres. This rapid and generalisable learning ability is
mostly due to the compositional structure of image patterns in the human brain,
which is less incorporated in medical image segmentation. In this paper, we
model the compositional components (i.e. patterns) of human anatomy as
learnable von-Mises-Fisher (vMF) kernels, which are robust to images collected
from different domains (e.g. clinical centres). The image features can be
decomposed to (or composed by) the components with the composing operations,
i.e. the vMF likelihoods. The vMF likelihoods tell how likely each anatomical
part is at each position of the image. Hence, the segmentation mask can be
predicted based on the vMF likelihoods. Moreover, with a reconstruction module,
unlabeled data can also be used to learn the vMF kernels and likelihoods by
recombining them to reconstruct the input image. Extensive experiments show
that the proposed vMFNet achieves improved generalisation performance on two
benchmarks, especially when annotations are limited. Code is publicly available
at: https://github.com/vios-s/vMFNet.
Related papers
- Do Vision Foundation Models Enhance Domain Generalization in Medical Image Segmentation? [10.20366295974822]
We introduce a novel decode head architecture, HQHSAM, which simply integrates elements from two state-of-the-art decoder heads, HSAM and HQSAM, to enhance segmentation performance.
Our experiments on multiple datasets, encompassing various anatomies and modalities, reveal that FMs, particularly with the HQHSAM decode head, improve domain generalization for medical image segmentation.
arXiv Detail & Related papers (2024-09-12T11:41:35Z) - MindFormer: Semantic Alignment of Multi-Subject fMRI for Brain Decoding [50.55024115943266]
We introduce a novel semantic alignment method of multi-subject fMRI signals using so-called MindFormer.
This model is specifically designed to generate fMRI-conditioned feature vectors that can be used for conditioning Stable Diffusion model for fMRI- to-image generation or large language model (LLM) for fMRI-to-text generation.
Our experimental results demonstrate that MindFormer generates semantically consistent images and text across different subjects.
arXiv Detail & Related papers (2024-05-28T00:36:25Z) - Teaching AI the Anatomy Behind the Scan: Addressing Anatomical Flaws in Medical Image Segmentation with Learnable Prior [34.54360931760496]
Key anatomical features, such as the number of organs, their shapes and relative positions, are crucial for building a robust multi-organ segmentation model.
We introduce a novel architecture called the Anatomy-Informed Network (AIC-Net)
AIC-Net incorporates a learnable input termed "Anatomical Prior", which can be adapted to patient-specific anatomy.
arXiv Detail & Related papers (2024-03-27T10:46:24Z) - SM2C: Boost the Semi-supervised Segmentation for Medical Image by using Meta Pseudo Labels and Mixed Images [13.971120210536995]
We introduce Scaling-up Mix with Multi-Class (SM2C) to improve the ability to learn semantic features within medical images.
By diversifying the shape of the segmentation objects and enriching the semantic information within each sample, the SM2C demonstrates its potential.
The proposed framework shows significant improvements over state-of-the-art counterparts.
arXiv Detail & Related papers (2024-03-24T04:39:40Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Compositionally Equivariant Representation Learning [22.741376970643973]
Humans can swiftly learn to identify important anatomy in medical images like MRI and CT scans.
This recognition capability easily generalises to new images from different medical facilities and to new tasks in different settings.
We study the utilisation of compositionality in learning more interpretable and generalisable representations for medical image segmentation.
arXiv Detail & Related papers (2023-06-13T14:06:55Z) - BrainCLIP: Bridging Brain and Visual-Linguistic Representation Via CLIP
for Generic Natural Visual Stimulus Decoding [51.911473457195555]
BrainCLIP is a task-agnostic fMRI-based brain decoding model.
It bridges the modality gap between brain activity, image, and text.
BrainCLIP can reconstruct visual stimuli with high semantic fidelity.
arXiv Detail & Related papers (2023-02-25T03:28:54Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Recurrent Mask Refinement for Few-Shot Medical Image Segmentation [15.775057485500348]
We propose a new framework for few-shot medical image segmentation based on prototypical networks.
Our innovation lies in the design of two key modules: 1) a context relation encoder (CRE) that uses correlation to capture local relation features between foreground and background regions.
Experiments on two abdomen CT datasets and an abdomen MRI dataset show the proposed method obtains substantial improvement over the state-of-the-art methods.
arXiv Detail & Related papers (2021-08-02T04:06:12Z) - Generalized Organ Segmentation by Imitating One-shot Reasoning using
Anatomical Correlation [55.1248480381153]
We propose OrganNet which learns a generalized organ concept from a set of annotated organ classes and then transfer this concept to unseen classes.
We show that OrganNet can effectively resist the wide variations in organ morphology and produce state-of-the-art results in one-shot segmentation task.
arXiv Detail & Related papers (2021-03-30T13:41:12Z) - Few-shot Medical Image Segmentation using a Global Correlation Network
with Discriminative Embedding [60.89561661441736]
We propose a novel method for few-shot medical image segmentation.
We construct our few-shot image segmentor using a deep convolutional network trained episodically.
We enhance discriminability of deep embedding to encourage clustering of the feature domains of the same class.
arXiv Detail & Related papers (2020-12-10T04:01:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.