$\texttt{NePhi}$: Neural Deformation Fields for Approximately Diffeomorphic Medical Image Registration
- URL: http://arxiv.org/abs/2309.07322v3
- Date: Fri, 27 Sep 2024 03:14:30 GMT
- Title: $\texttt{NePhi}$: Neural Deformation Fields for Approximately Diffeomorphic Medical Image Registration
- Authors: Lin Tian, Hastings Greer, Raúl San José Estépar, Roni Sengupta, Marc Niethammer,
- Abstract summary: NePhi represents deformations functionally, leading to great flexibility within the design space of memory consumption.
We show that NePhi can match the accuracy of voxel-based representations in a single-resolution registration setting.
For multi-resolution registration, our method matches the accuracy of current SOTA learning-based registration approaches with instance optimization.
- Score: 16.388101540950295
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This work proposes NePhi, a generalizable neural deformation model which results in approximately diffeomorphic transformations. In contrast to the predominant voxel-based transformation fields used in learning-based registration approaches, NePhi represents deformations functionally, leading to great flexibility within the design space of memory consumption during training and inference, inference time, registration accuracy, as well as transformation regularity. Specifically, NePhi 1) requires less memory compared to voxel-based learning approaches, 2) improves inference speed by predicting latent codes, compared to current existing neural deformation based registration approaches that \emph{only} rely on optimization, 3) improves accuracy via instance optimization, and 4) shows excellent deformation regularity which is highly desirable for medical image registration. We demonstrate the performance of NePhi on a 2D synthetic dataset as well as for real 3D medical image datasets (e.g., lungs and brains). Our results show that NePhi can match the accuracy of voxel-based representations in a single-resolution registration setting. For multi-resolution registration, our method matches the accuracy of current SOTA learning-based registration approaches with instance optimization while reducing memory requirements by a factor of five. Our code is available at https://github.com/uncbiag/NePhi.
Related papers
- GSMorph: Gradient Surgery for cine-MRI Cardiac Deformable Registration [62.41725951450803]
Learning-based deformable registration relies on weighted objective functions trading off registration accuracy and smoothness of the field.
We construct a registration model based on the gradient surgery mechanism, named GSMorph, to achieve a hyper parameter-free balance on multiple losses.
Our method is model-agnostic and can be merged into any deep registration network without introducing extra parameters or slowing down inference.
arXiv Detail & Related papers (2023-06-26T13:32:09Z) - Neural Priming for Sample-Efficient Adaptation [92.14357804106787]
We propose Neural Priming, a technique for adapting large pretrained models to distribution shifts and downstream tasks.
Neural Priming can be performed at test time, even for pretraining as large as LAION-2B.
arXiv Detail & Related papers (2023-06-16T21:53:16Z) - General Neural Gauge Fields [100.35916421218101]
We develop a learning framework to jointly optimize gauge transformations and neural fields.
We derive an information-invariant gauge transformation which allows to preserve scene information inherently and yield superior performance.
arXiv Detail & Related papers (2023-05-05T12:08:57Z) - Anatomy-aware and acquisition-agnostic joint registration with SynthMorph [6.017634371712142]
Affine image registration is a cornerstone of medical image analysis.
Deep-learning (DL) methods learn a function that maps an image pair to an output transform.
Most affine methods are agnostic to the anatomy the user wishes to align, meaning the registration will be inaccurate if algorithms consider all structures in the image.
We address these shortcomings with SynthMorph, a fast, symmetric, diffeomorphic, and easy-to-use DL tool for joint affine-deformable registration of any brain image.
arXiv Detail & Related papers (2023-01-26T18:59:33Z) - Fast-SNARF: A Fast Deformer for Articulated Neural Fields [92.68788512596254]
We propose a new articulation module for neural fields, Fast-SNARF, which finds accurate correspondences between canonical space and posed space.
Fast-SNARF is a drop-in replacement in to our previous work, SNARF, while significantly improving its computational efficiency.
Because learning of deformation maps is a crucial component in many 3D human avatar methods, we believe that this work represents a significant step towards the practical creation of 3D virtual humans.
arXiv Detail & Related papers (2022-11-28T17:55:34Z) - Adapting the Mean Teacher for keypoint-based lung registration under
geometric domain shifts [75.51482952586773]
deep neural networks generally require plenty of labeled training data and are vulnerable to domain shifts between training and test data.
We present a novel approach to geometric domain adaptation for image registration, adapting a model from a labeled source to an unlabeled target domain.
Our method consistently improves on the baseline model by 50%/47% while even matching the accuracy of models trained on target data.
arXiv Detail & Related papers (2022-07-01T12:16:42Z) - Medical Image Registration via Neural Fields [35.80302878742334]
We present a new neural net based image registration framework, called NIR (Neural Image Registration), which is based on optimization but utilizes deep neural nets to model deformations between image pairs.
Experiments on two 3D MR brain scan datasets demonstrate that NIR yields state-of-the-art performance in terms of both registration accuracy and regularity, while running significantly faster than traditional optimization-based methods.
arXiv Detail & Related papers (2022-06-07T08:43:31Z) - Affine Medical Image Registration with Coarse-to-Fine Vision Transformer [11.4219428942199]
We present a learning-based algorithm, Coarse-to-Fine Vision Transformer (C2FViT), for 3D affine medical image registration.
Our method is superior to the existing CNNs-based affine registration methods in terms of registration accuracy, robustness and generalizability.
arXiv Detail & Related papers (2022-03-29T03:18:43Z) - Deformable Image Registration using Neural ODEs [15.245085400790002]
We present a generic, fast, and accurate diffeomorphic image registration framework that leverages neural ordinary differential equations (NODEs)
Compared with traditional optimization-based methods, our framework reduces the running time from tens of minutes to tens of seconds.
Our experiments show that the registration results of our method outperform state-of-the-arts under various metrics.
arXiv Detail & Related papers (2021-08-07T12:54:17Z) - Deep Learning for Regularization Prediction in Diffeomorphic Image
Registration [8.781861951759948]
We introduce a novel framework that automatically determines the parameters controlling the smoothness of diffeomorphic transformations.
We develop a predictive model based on deep convolutional neural networks (CNN) that learns the mapping between pairwise images and the regularization parameter of image registration.
Experimental results show that our model not only predicts appropriate regularization parameters for image registration, but also improving the network training in terms of time and memory efficiency.
arXiv Detail & Related papers (2020-11-28T22:56:44Z) - DeepGMR: Learning Latent Gaussian Mixture Models for Registration [113.74060941036664]
Point cloud registration is a fundamental problem in 3D computer vision, graphics and robotics.
In this paper, we introduce Deep Gaussian Mixture Registration (DeepGMR), the first learning-based registration method.
Our proposed method shows favorable performance when compared with state-of-the-art geometry-based and learning-based registration methods.
arXiv Detail & Related papers (2020-08-20T17:25:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.