Learning Generalized Non-Rigid Multimodal Biomedical Image Registration
from Generic Point Set Data
- URL: http://arxiv.org/abs/2207.10994v1
- Date: Fri, 22 Jul 2022 10:27:54 GMT
- Title: Learning Generalized Non-Rigid Multimodal Biomedical Image Registration
from Generic Point Set Data
- Authors: Zachary MC Baum, Tamas Ungi, Christopher Schlenger, Yipeng Hu, Dean C
Barratt
- Abstract summary: Free Point Transformer (FPT) has been proposed as a data-driven, non-rigid point set registration approach using deep neural networks.
This work explores the generalizability of the FPT from well-curated non-medical data sets to medical imaging data sets.
- Score: 0.30544323433686693
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Free Point Transformer (FPT) has been proposed as a data-driven, non-rigid
point set registration approach using deep neural networks. As FPT does not
assume constraints based on point vicinity or correspondence, it may be trained
simply and in a flexible manner by minimizing an unsupervised loss based on the
Chamfer Distance. This makes FPT amenable to real-world medical imaging
applications where ground-truth deformations may be infeasible to obtain, or in
scenarios where only a varying degree of completeness in the point sets to be
aligned is available. To test the limit of the correspondence finding ability
of FPT and its dependency on training data sets, this work explores the
generalizability of the FPT from well-curated non-medical data sets to medical
imaging data sets. First, we train FPT on the ModelNet40 dataset to demonstrate
its effectiveness and the superior registration performance of FPT over
iterative and learning-based point set registration methods. Second, we
demonstrate superior performance in rigid and non-rigid registration and
robustness to missing data. Last, we highlight the interesting generalizability
of the ModelNet-trained FPT by registering reconstructed freehand ultrasound
scans of the spine and generic spine models without additional training,
whereby the average difference to the ground truth curvatures is 1.3 degrees,
across 13 patients.
Related papers
- Probing the Efficacy of Federated Parameter-Efficient Fine-Tuning of Vision Transformers for Medical Image Classification [16.070261684997362]
Fine-tuning pre-trained models for various downstream tasks is a critical problem in the medical imaging domain.
Large size of these models necessitates the use of parameter-efficient fine-tuning (PEFT) to reduce the communication burden in federated learning.
In this work, we investigate various federated PEFT strategies for adapting a Vision Transformer (ViT) model for medical image classification.
arXiv Detail & Related papers (2024-07-16T10:28:50Z) - Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - Diverse Data Augmentation with Diffusions for Effective Test-time Prompt
Tuning [73.75282761503581]
We propose DiffTPT, which leverages pre-trained diffusion models to generate diverse and informative new data.
Our experiments on test datasets with distribution shifts and unseen categories demonstrate that DiffTPT improves the zero-shot accuracy by an average of 5.13%.
arXiv Detail & Related papers (2023-08-11T09:36:31Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language
Models [107.05966685291067]
We propose test-time prompt tuning (TPT) to learn adaptive prompts on the fly with a single test sample.
TPT improves the zero-shot top-1 accuracy of CLIP by 3.6% on average.
In evaluating cross-dataset generalization with unseen categories, TPT performs on par with the state-of-the-art approaches that use additional training data.
arXiv Detail & Related papers (2022-09-15T17:55:11Z) - Label-Efficient Self-Supervised Federated Learning for Tackling Data
Heterogeneity in Medical Imaging [23.08596805950814]
We present a robust and label-efficient self-supervised FL framework for medical image analysis.
Specifically, we introduce a novel distributed self-supervised pre-training paradigm into the existing FL pipeline.
We show that our self-supervised FL algorithm generalizes well to out-of-distribution data and learns federated models more effectively in limited label scenarios.
arXiv Detail & Related papers (2022-05-17T18:33:43Z) - CPT: Colorful Prompt Tuning for Pre-trained Vision-Language Models [101.5066760592534]
We present Cross-modal Prompt Tuning (CPT), a novel paradigm for tuning Vision-Language Models (VL-PTMs)
CPT reformulates visual grounding into a fill-in-the-blank problem with color-based co-referential markers in image and text, maximally mitigating the gap.
Comprehensive experimental results show that prompt tuned VL-PTMs outperform their fine-tuned counterparts by a large margin.
arXiv Detail & Related papers (2021-09-24T08:07:29Z) - Real-time multimodal image registration with partial intraoperative
point-set data [0.5625255382226245]
Free Point Transformer (FPT) is a deep neural network architecture for non-rigid point-set registration.
In this paper, we demonstrate the application of FPT to non-rigid registration of prostate magnetic resonance (MR) imaging and sparsely-sampled transrectal ultrasound (TRUS) images.
arXiv Detail & Related papers (2021-09-10T10:21:31Z) - Differentially private federated deep learning for multi-site medical
image segmentation [56.30543374146002]
Collaborative machine learning techniques such as federated learning (FL) enable the training of models on effectively larger datasets without data transfer.
Recent initiatives have demonstrated that segmentation models trained with FL can achieve performance similar to locally trained models.
However, FL is not a fully privacy-preserving technique and privacy-centred attacks can disclose confidential patient data.
arXiv Detail & Related papers (2021-07-06T12:57:32Z) - Multimodality Biomedical Image Registration using Free Point Transformer
Networks [0.37501702548174964]
We describe a point-set registration algorithm based on a novel free point transformer (FPT) network.
FPT is constructed with a global feature extractor which accepts unordered source and target point-sets of variable size.
In a multimodal registration task using prostate MR and sparsely acquired ultrasound images, FPT yields comparable or improved results.
arXiv Detail & Related papers (2020-08-05T00:13:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.