Multimodality Biomedical Image Registration using Free Point Transformer
Networks
- URL: http://arxiv.org/abs/2008.01885v1
- Date: Wed, 5 Aug 2020 00:13:04 GMT
- Title: Multimodality Biomedical Image Registration using Free Point Transformer
Networks
- Authors: Zachary M. C. Baum, Yipeng Hu, Dean C. Barratt
- Abstract summary: We describe a point-set registration algorithm based on a novel free point transformer (FPT) network.
FPT is constructed with a global feature extractor which accepts unordered source and target point-sets of variable size.
In a multimodal registration task using prostate MR and sparsely acquired ultrasound images, FPT yields comparable or improved results.
- Score: 0.37501702548174964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We describe a point-set registration algorithm based on a novel free point
transformer (FPT) network, designed for points extracted from multimodal
biomedical images for registration tasks, such as those frequently encountered
in ultrasound-guided interventional procedures. FPT is constructed with a
global feature extractor which accepts unordered source and target point-sets
of variable size. The extracted features are conditioned by a shared multilayer
perceptron point transformer module to predict a displacement vector for each
source point, transforming it into the target space. The point transformer
module assumes no vicinity or smoothness in predicting spatial transformation
and, together with the global feature extractor, is trained in a data-driven
fashion with an unsupervised loss function. In a multimodal registration task
using prostate MR and sparsely acquired ultrasound images, FPT yields
comparable or improved results over other rigid and non-rigid registration
methods. This demonstrates the versatility of FPT to learn registration
directly from real, clinical training data and to generalize to a challenging
task, such as the interventional application presented.
Related papers
- Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - FMRT: Learning Accurate Feature Matching with Reconciliatory Transformer [29.95553680263075]
We propose Feature Matching with Reconciliatory Transformer (FMRT), a detector-free method that reconciles different features with multiple receptive fields adaptively.
FMRT yields extraordinary performance on multiple benchmarks, including pose estimation, visual localization, homography estimation, and image matching.
arXiv Detail & Related papers (2023-10-20T15:54:18Z) - RegFormer: An Efficient Projection-Aware Transformer Network for
Large-Scale Point Cloud Registration [73.69415797389195]
We propose an end-to-end transformer network (RegFormer) for large-scale point cloud alignment.
Specifically, a projection-aware hierarchical transformer is proposed to capture long-range dependencies and filter outliers.
Our transformer has linear complexity, which guarantees high efficiency even for large-scale scenes.
arXiv Detail & Related papers (2023-03-22T08:47:37Z) - DPCN++: Differentiable Phase Correlation Network for Versatile Pose
Registration [18.60311260250232]
We present a differentiable phase correlation solver that is globally convergent and correspondence-free.
We evaluate DCPN++ on a wide range of registration tasks taking different input modalities, including 2D bird's-eye view images, 3D object and scene measurements, and medical images.
arXiv Detail & Related papers (2022-06-12T10:00:34Z) - Dispensed Transformer Network for Unsupervised Domain Adaptation [21.256375606219073]
A novel unsupervised domain adaptation (UDA) method named dispensed Transformer network (DTNet) is introduced in this paper.
Our proposed network achieves the best performance in comparison with several state-of-the-art techniques.
arXiv Detail & Related papers (2021-10-28T08:27:44Z) - PnP-DETR: Towards Efficient Visual Analysis with Transformers [146.55679348493587]
Recently, DETR pioneered the solution vision tasks with transformers, it directly translates the image feature map into the object result.
Recent transformer-based image recognition model andTT show consistent efficiency gain.
arXiv Detail & Related papers (2021-09-15T01:10:30Z) - Real-time multimodal image registration with partial intraoperative
point-set data [0.5625255382226245]
Free Point Transformer (FPT) is a deep neural network architecture for non-rigid point-set registration.
In this paper, we demonstrate the application of FPT to non-rigid registration of prostate magnetic resonance (MR) imaging and sparsely-sampled transrectal ultrasound (TRUS) images.
arXiv Detail & Related papers (2021-09-10T10:21:31Z) - Rethinking Global Context in Crowd Counting [70.54184500538338]
A pure transformer is used to extract features with global information from overlapping image patches.
Inspired by classification, we add a context token to the input sequence, to facilitate information exchange with tokens corresponding to image patches.
arXiv Detail & Related papers (2021-05-23T12:44:27Z) - Semi-Supervised Domain Adaptation with Prototypical Alignment and
Consistency Learning [86.6929930921905]
This paper studies how much it can help address domain shifts if we further have a few target samples labeled.
To explore the full potential of landmarks, we incorporate a prototypical alignment (PA) module which calculates a target prototype for each class from the landmarks.
Specifically, we severely perturb the labeled images, making PA non-trivial to achieve and thus promoting model generalizability.
arXiv Detail & Related papers (2021-04-19T08:46:08Z) - End-to-End Trainable Multi-Instance Pose Estimation with Transformers [68.93512627479197]
We propose a new end-to-end trainable approach for multi-instance pose estimation by combining a convolutional neural network with a transformer.
Inspired by recent work on end-to-end trainable object detection with transformers, we use a transformer encoder-decoder architecture together with a bipartite matching scheme to directly regress the pose of all individuals in a given image.
Our model, called POse Estimation Transformer (POET), is trained using a novel set-based global loss that consists of a keypoint loss, a keypoint visibility loss, a center loss and a class loss.
arXiv Detail & Related papers (2021-03-22T18:19:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.