Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth
Manifold Triangulation
- URL: http://arxiv.org/abs/2108.05269v1
- Date: Wed, 11 Aug 2021 15:11:34 GMT
- Title: Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth
Manifold Triangulation
- Authors: Jianning Li, Antonio Pepe, Christina Gsaxner, Yuan Jin, Jan Egger
- Abstract summary: We propose that high-resolution images can be reconstructed in a coarse-to-fine fashion, where a deep learning algorithm is only responsible for generating a coarse representation of the image.
For producing the high-resolution outcome, we propose two novel methods: learned voxel rearrangement of the coarse output and hierarchical image synthesis.
Compared to the coarse output, the high-resolution counterpart allows for smooth surface triangulation, which can be 3D-printed in the highest possible quality.
- Score: 0.8968417883198374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical images, especially volumetric images, are of high resolution and
often exceed the capacity of standard desktop GPUs. As a result, most deep
learning-based medical image analysis tasks require the input images to be
downsampled, often substantially, before these can be fed to a neural network.
However, downsampling can lead to a loss of image quality, which is undesirable
especially in reconstruction tasks, where the fine geometric details need to be
preserved. In this paper, we propose that high-resolution images can be
reconstructed in a coarse-to-fine fashion, where a deep learning algorithm is
only responsible for generating a coarse representation of the image, which
consumes moderate GPU memory. For producing the high-resolution outcome, we
propose two novel methods: learned voxel rearrangement of the coarse output and
hierarchical image synthesis. Compared to the coarse output, the
high-resolution counterpart allows for smooth surface triangulation, which can
be 3D-printed in the highest possible quality. Experiments of this paper are
carried out on the dataset of AutoImplant 2021
(https://autoimplant2021.grand-challenge.org/), a MICCAI challenge on cranial
implant design. The dataset contains high-resolution skulls that can be viewed
as 2D manifolds embedded in a 3D space. Codes associated with this study can be
accessed at https://github.com/Jianningli/voxel_rearrangement.
Related papers
- SCube: Instant Large-Scale Scene Reconstruction using VoxSplats [55.383993296042526]
We present SCube, a novel method for reconstructing large-scale 3D scenes (geometry, appearance, and semantics) from a sparse set of posed images.
Our method encodes reconstructed scenes using a novel representation VoxSplat, which is a set of 3D Gaussians supported on a high-resolution sparse-voxel scaffold.
arXiv Detail & Related papers (2024-10-26T00:52:46Z) - μ-Net: A Deep Learning-Based Architecture for μ-CT Segmentation [2.012378666405002]
X-ray computed microtomography (mu-CT) is a non-destructive technique that can generate high-resolution 3D images of the internal anatomy of medical and biological samples.
extracting relevant information from 3D images requires semantic segmentation of the regions of interest.
We propose a novel framework that uses a convolutional neural network (CNN) to automatically segment the full morphology of the heart of Carassius auratus.
arXiv Detail & Related papers (2024-06-24T15:29:08Z) - Geometry-Aware Attenuation Learning for Sparse-View CBCT Reconstruction [53.93674177236367]
Cone Beam Computed Tomography (CBCT) plays a vital role in clinical imaging.
Traditional methods typically require hundreds of 2D X-ray projections to reconstruct a high-quality 3D CBCT image.
This has led to a growing interest in sparse-view CBCT reconstruction to reduce radiation doses.
We introduce a novel geometry-aware encoder-decoder framework to solve this problem.
arXiv Detail & Related papers (2023-03-26T14:38:42Z) - Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with
Implicit Neural Representation [3.8215162658168524]
Oral-3Dv2 is a non-adversarial-learning-based model in 3D radiology reconstruction from a single panoramic X-ray image.
Our model learns to represent the 3D oral structure in an implicit way by mapping 2D coordinates into density values of voxels in the 3D space.
To the best of our knowledge, this is the first work of a non-adversarial-learning-based model in 3D radiology reconstruction from a single panoramic X-ray image.
arXiv Detail & Related papers (2023-03-21T18:17:27Z) - Decoding natural image stimuli from fMRI data with a surface-based
convolutional network [1.8352113484137629]
We propose a novel approach to decode visual stimuli with high semantic fidelity and rich fine-grained detail.
Our proposed method achieves state-of-the-art semantic fidelity, while yielding good fine-grained similarity with the ground-truth stimulus.
arXiv Detail & Related papers (2022-12-05T16:47:19Z) - A Self-Supervised Approach to Reconstruction in Sparse X-Ray Computed
Tomography [1.0806206850043696]
This work develops and validates a self-supervised probabilistic deep learning technique, the physics-informed variational autoencoder.
Deep neural networks have been used to transform sparse 2-D projection measurements to a 3-D reconstruction by training on a dataset of known similar objects.
High-quality reconstructions cannot be generated without deep learning, and the deep neural network cannot be learned without the reconstructions.
arXiv Detail & Related papers (2022-10-30T02:33:45Z) - OSTeC: One-Shot Texture Completion [86.23018402732748]
We propose an unsupervised approach for one-shot 3D facial texture completion.
The proposed approach rotates an input image in 3D and fill-in the unseen regions by reconstructing the rotated image in a 2D face generator.
We frontalize the target image by projecting the completed texture into the generator.
arXiv Detail & Related papers (2020-12-30T23:53:26Z) - Learning Deformable Tetrahedral Meshes for 3D Reconstruction [78.0514377738632]
3D shape representations that accommodate learning-based 3D reconstruction are an open problem in machine learning and computer graphics.
Previous work on neural 3D reconstruction demonstrated benefits, but also limitations, of point cloud, voxel, surface mesh, and implicit function representations.
We introduce Deformable Tetrahedral Meshes (DefTet) as a particular parameterization that utilizes volumetric tetrahedral meshes for the reconstruction problem.
arXiv Detail & Related papers (2020-11-03T02:57:01Z) - Memory-efficient GAN-based Domain Translation of High Resolution 3D
Medical Images [0.15092198588928965]
Generative adversarial networks (GANs) are rarely applied on 3D medical images of large size.
The present work proposes a multi-scale patch-based GAN approach for establishing unpaired domain translation.
The evaluation of the domain translation scenarios is performed on brain MRIs of size 155x240x240 and thorax CTs of size up to 512x512x512.
arXiv Detail & Related papers (2020-10-06T08:43:27Z) - Hierarchical Amortized Training for Memory-efficient High Resolution 3D
GAN [52.851990439671475]
We propose a novel end-to-end GAN architecture that can generate high-resolution 3D images.
We achieve this goal by using different configurations between training and inference.
Experiments on 3D thorax CT and brain MRI demonstrate that our approach outperforms state of the art in image generation.
arXiv Detail & Related papers (2020-08-05T02:33:04Z) - Pix2Vox++: Multi-scale Context-aware 3D Object Reconstruction from
Single and Multiple Images [56.652027072552606]
We propose a novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++.
By using a well-designed encoder-decoder, it generates a coarse 3D volume from each input image.
A multi-scale context-aware fusion module is then introduced to adaptively select high-quality reconstructions for different parts from all coarse 3D volumes to obtain a fused 3D volume.
arXiv Detail & Related papers (2020-06-22T13:48:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.