LDDMM meets GANs: Generative Adversarial Networks for diffeomorphic
registration
- URL: http://arxiv.org/abs/2111.12544v1
- Date: Wed, 24 Nov 2021 15:26:16 GMT
- Title: LDDMM meets GANs: Generative Adversarial Networks for diffeomorphic
registration
- Authors: Ubaldo Ramon, Monica Hernandez, and Elvira Mayordomo
- Abstract summary: We propose an adversarial learning method for pairs of 3D mono-modal images based on Generative Adrial Networks.
We have successfully implemented two models with the stationary and the EPDiff-constrained non-stationary parameterizations of diffeomorphisms.
Our method has shown similar results to model-based methods with a computational time under one second.
- Score: 1.2599533416395767
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The purpose of this work is to contribute to the state of the art of
deep-learning methods for diffeomorphic registration. We propose an adversarial
learning LDDMM method for pairs of 3D mono-modal images based on Generative
Adversarial Networks. The method is inspired by the recent literature for
deformable image registration with adversarial learning. We combine the best
performing generative, discriminative, and adversarial ingredients from the
state of the art within the LDDMM paradigm. We have successfully implemented
two models with the stationary and the EPDiff-constrained non-stationary
parameterizations of diffeomorphisms. Our unsupervised and data-hungry approach
has shown a competitive performance with respect to a benchmark supervised and
rich-data approach. In addition, our method has shown similar results to
model-based methods with a computational time under one second.
Related papers
- Unified Speech Recognition: A Single Model for Auditory, Visual, and Audiovisual Inputs [73.74375912785689]
This paper proposes unified training strategies for speech recognition systems.
We demonstrate that training a single model for all three tasks enhances VSR and AVSR performance.
We also introduce a greedy pseudo-labelling approach to more effectively leverage unlabelled samples.
arXiv Detail & Related papers (2024-11-04T16:46:53Z) - Pose-Guided Self-Training with Two-Stage Clustering for Unsupervised Landmark Discovery [17.455841673719625]
Unsupervised landmarks discovery (ULD) for an object category is a challenging computer vision problem.
In pursuit of developing a robust ULD framework, we explore the potential of a recent paradigm of self-supervised learning algorithms, known as diffusion models.
Our approach consistently outperforms state-of-the-art methods on four challenging benchmarks AFLW, MAFL, CatHeads and LS3D by significant margins.
arXiv Detail & Related papers (2024-03-24T15:24:04Z) - Importance of Aligning Training Strategy with Evaluation for Diffusion
Models in 3D Multiclass Segmentation [8.649603931882227]
We studied the DDPM-based segmentation model for 3D multiclass segmentation on two large multiclass data sets.
We observed that the difference between training and test methods led to inferior performance for existing DDPM methods.
To mitigate the inconsistency, we proposed a recycling method which generated corrupted masks based on the model's prediction at a previous time step.
arXiv Detail & Related papers (2023-03-10T16:30:09Z) - Gait Recognition in the Wild with Multi-hop Temporal Switch [81.35245014397759]
gait recognition in the wild is a more practical problem that has attracted the attention of the community of multimedia and computer vision.
This paper presents a novel multi-hop temporal switch method to achieve effective temporal modeling of gait patterns in real-world scenes.
arXiv Detail & Related papers (2022-09-01T10:46:09Z) - Learning Gaussian Graphical Models with Latent Confounders [74.72998362041088]
We compare and contrast two strategies for inference in graphical models with latent confounders.
While these two approaches have similar goals, they are motivated by different assumptions about confounding.
We propose a new method, which combines the strengths of these two approaches.
arXiv Detail & Related papers (2021-05-14T00:53:03Z) - IMAGINE: Image Synthesis by Image-Guided Model Inversion [79.4691654458141]
We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images.
We leverage the knowledge of image semantics from a pre-trained classifier to achieve plausible generations.
IMAGINE enables the synthesis procedure to simultaneously 1) enforce semantic specificity constraints during the synthesis, 2) produce realistic images without generator training, and 3) give users intuitive control over the generation process.
arXiv Detail & Related papers (2021-04-13T02:00:24Z) - DeepGMR: Learning Latent Gaussian Mixture Models for Registration [113.74060941036664]
Point cloud registration is a fundamental problem in 3D computer vision, graphics and robotics.
In this paper, we introduce Deep Gaussian Mixture Registration (DeepGMR), the first learning-based registration method.
Our proposed method shows favorable performance when compared with state-of-the-art geometry-based and learning-based registration methods.
arXiv Detail & Related papers (2020-08-20T17:25:16Z) - Learning Deformable Image Registration from Optimization: Perspective,
Modules, Bilevel Training and Beyond [62.730497582218284]
We develop a new deep learning based framework to optimize a diffeomorphic model via multi-scale propagation.
We conduct two groups of image registration experiments on 3D volume datasets including image-to-atlas registration on brain MRI data and image-to-image registration on liver CT data.
arXiv Detail & Related papers (2020-04-30T03:23:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.