Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth
Manifold Triangulation
- URL: http://arxiv.org/abs/2108.05269v1
- Date: Wed, 11 Aug 2021 15:11:34 GMT
- Title: Learning to Rearrange Voxels in Binary Segmentation Masks for Smooth
Manifold Triangulation
- Authors: Jianning Li, Antonio Pepe, Christina Gsaxner, Yuan Jin, Jan Egger
- Abstract summary: We propose that high-resolution images can be reconstructed in a coarse-to-fine fashion, where a deep learning algorithm is only responsible for generating a coarse representation of the image.
For producing the high-resolution outcome, we propose two novel methods: learned voxel rearrangement of the coarse output and hierarchical image synthesis.
Compared to the coarse output, the high-resolution counterpart allows for smooth surface triangulation, which can be 3D-printed in the highest possible quality.
- Score: 0.8968417883198374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical images, especially volumetric images, are of high resolution and
often exceed the capacity of standard desktop GPUs. As a result, most deep
learning-based medical image analysis tasks require the input images to be
downsampled, often substantially, before these can be fed to a neural network.
However, downsampling can lead to a loss of image quality, which is undesirable
especially in reconstruction tasks, where the fine geometric details need to be
preserved. In this paper, we propose that high-resolution images can be
reconstructed in a coarse-to-fine fashion, where a deep learning algorithm is
only responsible for generating a coarse representation of the image, which
consumes moderate GPU memory. For producing the high-resolution outcome, we
propose two novel methods: learned voxel rearrangement of the coarse output and
hierarchical image synthesis. Compared to the coarse output, the
high-resolution counterpart allows for smooth surface triangulation, which can
be 3D-printed in the highest possible quality. Experiments of this paper are
carried out on the dataset of AutoImplant 2021
(https://autoimplant2021.grand-challenge.org/), a MICCAI challenge on cranial
implant design. The dataset contains high-resolution skulls that can be viewed
as 2D manifolds embedded in a 3D space. Codes associated with this study can be
accessed at https://github.com/Jianningli/voxel_rearrangement.
Related papers
- Multimodal Visual Surrogate Compression for Alzheimer's Disease Classification [69.87877580725768]
Multimodal Visual Surrogate Compression (MVSC) learns to compress and adapt large 3D sMRI volumes into compact 2D features.<n>MVSC has two key components: a Volume Context that captures global cross-slice context under textual guidance, and an Adaptive Slice Fusion module that aggregates slice-level information in a text-enhanced, patch-wise manner.
arXiv Detail & Related papers (2026-01-29T13:05:46Z) - TomoGraphView: 3D Medical Image Classification with Omnidirectional Slice Representations and Graph Neural Networks [2.2906925991630085]
3D medical image classification remains a challenging task due to the complex spatial relationships and long-range dependencies inherent in accessible data.<n>Recent studies have highlighted the potential of 2D vision foundation models, originally trained on natural images, as powerful feature extractors for medical image analysis.<n>We propose TomoGraphView, a novel framework that integrates omnidirectional volume slicing with spherical graph-based feature aggregation.
arXiv Detail & Related papers (2025-11-12T16:30:34Z) - NUDF: Neural Unsigned Distance Fields for high resolution 3D medical image segmentation [0.13431733228151765]
We propose to learn a Neural Unsigned Distance Field (NUDF) directly from the image.
We evaluate our method on the task of left atrial appendage (LAA) segmentation from Computed Tomography (CT) images.
We are able to predict 3D mesh models that capture the details of the LAA and achieve accuracy in the order of the voxel spacing in the CT images.
arXiv Detail & Related papers (2025-04-25T13:32:16Z) - SCube: Instant Large-Scale Scene Reconstruction using VoxSplats [55.383993296042526]
We present SCube, a novel method for reconstructing large-scale 3D scenes (geometry, appearance, and semantics) from a sparse set of posed images.
Our method encodes reconstructed scenes using a novel representation VoxSplat, which is a set of 3D Gaussians supported on a high-resolution sparse-voxel scaffold.
arXiv Detail & Related papers (2024-10-26T00:52:46Z) - μ-Net: A Deep Learning-Based Architecture for μ-CT Segmentation [2.012378666405002]
X-ray computed microtomography (mu-CT) is a non-destructive technique that can generate high-resolution 3D images of the internal anatomy of medical and biological samples.
extracting relevant information from 3D images requires semantic segmentation of the regions of interest.
We propose a novel framework that uses a convolutional neural network (CNN) to automatically segment the full morphology of the heart of Carassius auratus.
arXiv Detail & Related papers (2024-06-24T15:29:08Z) - Geometry-Aware Attenuation Learning for Sparse-View CBCT Reconstruction [53.93674177236367]
Cone Beam Computed Tomography (CBCT) plays a vital role in clinical imaging.
Traditional methods typically require hundreds of 2D X-ray projections to reconstruct a high-quality 3D CBCT image.
This has led to a growing interest in sparse-view CBCT reconstruction to reduce radiation doses.
We introduce a novel geometry-aware encoder-decoder framework to solve this problem.
arXiv Detail & Related papers (2023-03-26T14:38:42Z) - Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with
Implicit Neural Representation [3.8215162658168524]
Oral-3Dv2 is a non-adversarial-learning-based model in 3D radiology reconstruction from a single panoramic X-ray image.
Our model learns to represent the 3D oral structure in an implicit way by mapping 2D coordinates into density values of voxels in the 3D space.
To the best of our knowledge, this is the first work of a non-adversarial-learning-based model in 3D radiology reconstruction from a single panoramic X-ray image.
arXiv Detail & Related papers (2023-03-21T18:17:27Z) - Decoding natural image stimuli from fMRI data with a surface-based
convolutional network [1.8352113484137629]
We propose a novel approach to decode visual stimuli with high semantic fidelity and rich fine-grained detail.
Our proposed method achieves state-of-the-art semantic fidelity, while yielding good fine-grained similarity with the ground-truth stimulus.
arXiv Detail & Related papers (2022-12-05T16:47:19Z) - A Self-Supervised Approach to Reconstruction in Sparse X-Ray Computed
Tomography [1.0806206850043696]
This work develops and validates a self-supervised probabilistic deep learning technique, the physics-informed variational autoencoder.
Deep neural networks have been used to transform sparse 2-D projection measurements to a 3-D reconstruction by training on a dataset of known similar objects.
High-quality reconstructions cannot be generated without deep learning, and the deep neural network cannot be learned without the reconstructions.
arXiv Detail & Related papers (2022-10-30T02:33:45Z) - OSTeC: One-Shot Texture Completion [86.23018402732748]
We propose an unsupervised approach for one-shot 3D facial texture completion.
The proposed approach rotates an input image in 3D and fill-in the unseen regions by reconstructing the rotated image in a 2D face generator.
We frontalize the target image by projecting the completed texture into the generator.
arXiv Detail & Related papers (2020-12-30T23:53:26Z) - Learning Deformable Tetrahedral Meshes for 3D Reconstruction [78.0514377738632]
3D shape representations that accommodate learning-based 3D reconstruction are an open problem in machine learning and computer graphics.
Previous work on neural 3D reconstruction demonstrated benefits, but also limitations, of point cloud, voxel, surface mesh, and implicit function representations.
We introduce Deformable Tetrahedral Meshes (DefTet) as a particular parameterization that utilizes volumetric tetrahedral meshes for the reconstruction problem.
arXiv Detail & Related papers (2020-11-03T02:57:01Z) - Memory-efficient GAN-based Domain Translation of High Resolution 3D
Medical Images [0.15092198588928965]
Generative adversarial networks (GANs) are rarely applied on 3D medical images of large size.
The present work proposes a multi-scale patch-based GAN approach for establishing unpaired domain translation.
The evaluation of the domain translation scenarios is performed on brain MRIs of size 155x240x240 and thorax CTs of size up to 512x512x512.
arXiv Detail & Related papers (2020-10-06T08:43:27Z) - Hierarchical Amortized Training for Memory-efficient High Resolution 3D
GAN [52.851990439671475]
We propose a novel end-to-end GAN architecture that can generate high-resolution 3D images.
We achieve this goal by using different configurations between training and inference.
Experiments on 3D thorax CT and brain MRI demonstrate that our approach outperforms state of the art in image generation.
arXiv Detail & Related papers (2020-08-05T02:33:04Z) - Pix2Vox++: Multi-scale Context-aware 3D Object Reconstruction from
Single and Multiple Images [56.652027072552606]
We propose a novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++.
By using a well-designed encoder-decoder, it generates a coarse 3D volume from each input image.
A multi-scale context-aware fusion module is then introduced to adaptively select high-quality reconstructions for different parts from all coarse 3D volumes to obtain a fused 3D volume.
arXiv Detail & Related papers (2020-06-22T13:48:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.