Memory-efficient GAN-based Domain Translation of High Resolution 3D
Medical Images
- URL: http://arxiv.org/abs/2010.03396v1
- Date: Tue, 6 Oct 2020 08:43:27 GMT
- Title: Memory-efficient GAN-based Domain Translation of High Resolution 3D
Medical Images
- Authors: Hristina Uzunova, Jan Ehrhardt, Heinz Handels
- Abstract summary: Generative adversarial networks (GANs) are rarely applied on 3D medical images of large size.
The present work proposes a multi-scale patch-based GAN approach for establishing unpaired domain translation.
The evaluation of the domain translation scenarios is performed on brain MRIs of size 155x240x240 and thorax CTs of size up to 512x512x512.
- Score: 0.15092198588928965
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial networks (GANs) are currently rarely applied on 3D
medical images of large size, due to their immense computational demand. The
present work proposes a multi-scale patch-based GAN approach for establishing
unpaired domain translation by generating 3D medical image volumes of high
resolution in a memory-efficient way. The key idea to enable memory-efficient
image generation is to first generate a low-resolution version of the image
followed by the generation of patches of constant sizes but successively
growing resolutions. To avoid patch artifacts and incorporate global
information, the patch generation is conditioned on patches from previous
resolution scales. Those multi-scale GANs are trained to generate realistically
looking images from image sketches in order to perform an unpaired domain
translation. This allows to preserve the topology of the test data and generate
the appearance of the training domain data. The evaluation of the domain
translation scenarios is performed on brain MRIs of size 155x240x240 and thorax
CTs of size up to 512x512x512. Compared to common patch-based approaches, the
multi-resolution scheme enables better image quality and prevents patch
artifacts. Also, it ensures constant GPU memory demand independent from the
image size, allowing for the generation of arbitrarily large images.
Related papers
- Any-Size-Diffusion: Toward Efficient Text-Driven Synthesis for Any-Size
HD Images [56.17404812357676]
Stable diffusion, a generative model used in text-to-image synthesis, frequently encounters composition problems when generating images of varying sizes.
We propose a two-stage pipeline named Any-Size-Diffusion (ASD), designed to efficiently generate well-composed images of any size.
We show that ASD can produce well-structured images of arbitrary sizes, cutting down the inference time by 2x compared to the traditional tiled algorithm.
arXiv Detail & Related papers (2023-08-31T09:27:56Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Memory-Efficient 3D Denoising Diffusion Models for Medical Image Processing [0.9424565541639366]
We present a number of ways to reduce the resource consumption for 3D diffusion models.
The main contribution of this paper is the memory-efficient patch-based diffusion model.
While the proposed diffusion model can be applied to any image generation tasks, we evaluate the method on the tumor segmentation task of the BraTS 2020 dataset.
arXiv Detail & Related papers (2023-03-27T15:10:19Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - HIPA: Hierarchical Patch Transformer for Single Image Super Resolution [62.7081074931892]
This paper presents HIPA, a novel Transformer architecture that progressively recovers the high resolution image using a hierarchical patch partition.
We build a cascaded model that processes an input image in multiple stages, where we start with tokens with small patch sizes and gradually merge to the full resolution.
Such a hierarchical patch mechanism not only explicitly enables feature aggregation at multiple resolutions but also adaptively learns patch-aware features for different image regions.
arXiv Detail & Related papers (2022-03-19T05:09:34Z) - MICDIR: Multi-scale Inverse-consistent Deformable Image Registration
using UNetMSS with Self-Constructing Graph Latent [0.0]
This paper extends the Voxelmorph approach in three different ways.
To improve the performance in case of small as well as large deformations, supervision of the model at different resolutions has been integrated using a multi-scale UNet.
On the task of registration of brain MRIs, the proposed method achieved significant improvements over ANTs and VoxelMorph.
arXiv Detail & Related papers (2022-03-08T18:07:47Z) - Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth
Manifold Triangulation [0.8968417883198374]
We propose that high-resolution images can be reconstructed in a coarse-to-fine fashion, where a deep learning algorithm is only responsible for generating a coarse representation of the image.
For producing the high-resolution outcome, we propose two novel methods: learned voxel rearrangement of the coarse output and hierarchical image synthesis.
Compared to the coarse output, the high-resolution counterpart allows for smooth surface triangulation, which can be 3D-printed in the highest possible quality.
arXiv Detail & Related papers (2021-08-11T15:11:34Z) - SOUP-GAN: Super-Resolution MRI Using Generative Adversarial Networks [9.201328999176402]
We propose a framework called SOUP-GAN: Super-resolution Optimized Using Perceptual-tuned Generative Adversarial Network (GAN)
Our model shows promise as a novel 3D SR technique, providing potential applications in both clinical and research settings.
arXiv Detail & Related papers (2021-06-04T16:59:23Z) - InfinityGAN: Towards Infinite-Resolution Image Synthesis [92.40782797030977]
We present InfinityGAN, a method to generate arbitrary-resolution images.
We show how it trains and infers patch-by-patch seamlessly with low computational resources.
arXiv Detail & Related papers (2021-04-08T17:59:30Z) - Multi-Texture GAN: Exploring the Multi-Scale Texture Translation for
Brain MR Images [1.9163481966968943]
A significant percentage of existing algorithms cannot explicitly exploit and preserve texture details from target scanners.
In this paper, we design a multi-scale texture transfer to enrich the reconstruction images with more details.
Our method achieves superior results in inter-protocol or inter-scanner translation over state-of-the-art methods.
arXiv Detail & Related papers (2021-02-14T19:14:06Z) - Hierarchical Amortized Training for Memory-efficient High Resolution 3D
GAN [52.851990439671475]
We propose a novel end-to-end GAN architecture that can generate high-resolution 3D images.
We achieve this goal by using different configurations between training and inference.
Experiments on 3D thorax CT and brain MRI demonstrate that our approach outperforms state of the art in image generation.
arXiv Detail & Related papers (2020-08-05T02:33:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.