Inflating 2D Convolution Weights for Efficient Generation of 3D Medical
Images
- URL: http://arxiv.org/abs/2208.03934v3
- Date: Tue, 5 Dec 2023 23:59:59 GMT
- Title: Inflating 2D Convolution Weights for Efficient Generation of 3D Medical
Images
- Authors: Yanbin Liu, Girish Dwivedi, Farid Boussaid, Frank Sanfilippo, Makoto
Yamada, and Mohammed Bennamoun
- Abstract summary: Two problems prevent effective training of a 3D medical generative model: 3D medical images are expensive to acquire and annotate, and a large number of parameters are involved in 3D convolution.
We propose a novel GAN model called 3D Split&Shuffle-GAN.
We show that our method leads to improved 3D image generation quality with significantly fewer parameters.
- Score: 35.849240945334
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The generation of three-dimensional (3D) medical images has great application
potential since it takes into account the 3D anatomical structure. Two problems
prevent effective training of a 3D medical generative model: (1) 3D medical
images are expensive to acquire and annotate, resulting in an insufficient
number of training images, and (2) a large number of parameters are involved in
3D convolution.
Methods: We propose a novel GAN model called 3D Split&Shuffle-GAN. To address
the 3D data scarcity issue, we first pre-train a two-dimensional (2D) GAN model
using abundant image slices and inflate the 2D convolution weights to improve
the initialization of the 3D GAN. Novel 3D network architectures are proposed
for both the generator and discriminator of the GAN model to significantly
reduce the number of parameters while maintaining the quality of image
generation. Several weight inflation strategies and parameter-efficient 3D
architectures are investigated.
Results: Experiments on both heart (Stanford AIMI Coronary Calcium) and brain
(Alzheimer's Disease Neuroimaging Initiative) datasets show that our method
leads to improved 3D image generation quality (14.7 improvements on Fr\'echet
inception distance) with significantly fewer parameters (only 48.5% of the
baseline method).
Conclusions: We built a parameter-efficient 3D medical image generation
model. Due to the efficiency and effectiveness, it has the potential to
generate high-quality 3D brain and heart images for real use cases.
Related papers
- E3D-GPT: Enhanced 3D Visual Foundation for Medical Vision-Language Model [23.56751925900571]
The development of 3D medical vision-language models holds significant potential for disease diagnosis and patient treatment.
We utilize self-supervised learning to construct a 3D visual foundation model for extracting 3D visual features.
We apply 3D spatial convolutions to aggregate and project high-level image features, reducing computational complexity.
Our model demonstrates superior performance compared to existing methods in report generation, visual question answering, and disease diagnosis.
arXiv Detail & Related papers (2024-10-18T06:31:40Z) - Generative Enhancement for 3D Medical Images [74.17066529847546]
We propose GEM-3D, a novel generative approach to the synthesis of 3D medical images.
Our method begins with a 2D slice, noted as the informed slice to serve the patient prior, and propagates the generation process using a 3D segmentation mask.
By decomposing the 3D medical images into masks and patient prior information, GEM-3D offers a flexible yet effective solution for generating versatile 3D images.
arXiv Detail & Related papers (2024-03-19T15:57:04Z) - IM-3D: Iterative Multiview Diffusion and Reconstruction for High-Quality
3D Generation [96.32684334038278]
In this paper, we explore the design space of text-to-3D models.
We significantly improve multi-view generation by considering video instead of image generators.
Our new method, IM-3D, reduces the number of evaluations of the 2D generator network 10-100x.
arXiv Detail & Related papers (2024-02-13T18:59:51Z) - AG3D: Learning to Generate 3D Avatars from 2D Image Collections [96.28021214088746]
We propose a new adversarial generative model of realistic 3D people from 2D images.
Our method captures shape and deformation of the body and loose clothing by adopting a holistic 3D generator.
We experimentally find that our method outperforms previous 3D- and articulation-aware methods in terms of geometry and appearance.
arXiv Detail & Related papers (2023-05-03T17:56:24Z) - NeRF-GAN Distillation for Efficient 3D-Aware Generation with
Convolutions [97.27105725738016]
integration of Neural Radiance Fields (NeRFs) and generative models, such as Generative Adversarial Networks (GANs) has transformed 3D-aware generation from single-view images.
We propose a simple and effective method, based on re-using the well-disentangled latent space of a pre-trained NeRF-GAN in a pose-conditioned convolutional network to directly generate 3D-consistent images corresponding to the underlying 3D representations.
arXiv Detail & Related papers (2023-03-22T18:59:48Z) - Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models [52.529394863331326]
We propose a novel approach using two perpendicular pre-trained 2D diffusion models to solve the 3D inverse problem.
Our method is highly effective for 3D medical image reconstruction tasks, including MRI Z-axis super-resolution, compressed sensing MRI, and sparse-view CT.
arXiv Detail & Related papers (2023-03-15T08:28:06Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.