3D-OOCS: Learning Prostate Segmentation with Inductive Bias
- URL: http://arxiv.org/abs/2110.15664v1
- Date: Fri, 29 Oct 2021 10:14:56 GMT
- Title: 3D-OOCS: Learning Prostate Segmentation with Inductive Bias
- Authors: Shrajan Bhandary, Zahra Babaiee, Dejan Kostyszyn, Tobias Fechter,
Constantinos Zamboglou, Anca Grosu, Radu Grosu
- Abstract summary: We introduce OOCS-enhanced networks, a novel architecture inspired by the innate nature of visual processing in the vertebrates.
With different 3D U-Net variants as the base, we add two 3D residual components to the second encoder blocks: on and off center-surround.
OOCS helps 3D U-Nets to scrutinise and delineate anatomical structures present in 3D images with increased accuracy.
- Score: 5.907824204733372
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite the great success of convolutional neural networks (CNN) in 3D
medical image segmentation tasks, the methods currently in use are still not
robust enough to the different protocols utilized by different scanners, and to
the variety of image properties or artefacts they produce. To this end, we
introduce OOCS-enhanced networks, a novel architecture inspired by the innate
nature of visual processing in the vertebrates. With different 3D U-Net
variants as the base, we add two 3D residual components to the second encoder
blocks: on and off center-surround (OOCS). They generalise the ganglion
pathways in the retina to a 3D setting. The use of 2D-OOCS in any standard CNN
network complements the feedforward framework with sharp edge-detection
inductive biases. The use of 3D-OOCS also helps 3D U-Nets to scrutinise and
delineate anatomical structures present in 3D images with increased accuracy.We
compared the state-of-the-art 3D U-Nets with their 3D-OOCS extensions and
showed the superior accuracy and robustness of the latter in automatic prostate
segmentation from 3D Magnetic Resonance Images (MRIs). For a fair comparison,
we trained and tested all the investigated 3D U-Nets with the same pipeline,
including automatic hyperparameter optimisation and data augmentation.
Related papers
- 3D-Adapter: Geometry-Consistent Multi-View Diffusion for High-Quality 3D Generation [45.218605449572586]
3D-Adapter is a plug-in module designed to infuse 3D geometry awareness into pretrained image diffusion models.
We show that 3D-Adapter greatly enhances the geometry quality of text-to-multi-view models such as Instant3D and Zero123++.
We also showcase the broad application potential of 3D-Adapter by presenting high quality results in text-to-3D, image-to-3D, text-to-texture, and text-to-avatar tasks.
arXiv Detail & Related papers (2024-10-24T17:59:30Z) - SE3D: A Framework For Saliency Method Evaluation In 3D Imaging [4.090991964172346]
3D Convolutional Neural Networks (3D CNNs) are able to process LIDAR, MRI, and CT scans, with significant implications for fields such as autonomous driving and medical imaging.
Despite recent advances in Explainable Artificial Intelligence, little effort has been devoted to explaining 3D CNNs.
We propose SE3D: a framework for Saliency method Evaluation in 3D imaging.
arXiv Detail & Related papers (2024-05-23T13:55:11Z) - IB-U-Nets: Improving medical image segmentation tasks with 3D Inductive
Biased kernels [8.361151913935776]
We introduce IB-U-Nets, a novel architecture with inductive bias, inspired by the visual processing in vertebrates.
With the 3D U-Net as the base, we add two 3D residual components to the second encoder blocks. They provide an inductive bias, helping U-Nets to segment anatomical structures from 3D images with increased robustness and accuracy.
Our results demonstrate the superior robustness and accuracy of IB-U-Nets, especially on small datasets, as is typically the case in medical-image analysis.
arXiv Detail & Related papers (2022-10-28T07:12:15Z) - Neural Feature Fusion Fields: 3D Distillation of Self-Supervised 2D
Image Representations [92.88108411154255]
We present a method that improves dense 2D image feature extractors when the latter are applied to the analysis of multiple images reconstructible as a 3D scene.
We show that our method not only enables semantic understanding in the context of scene-specific neural fields without the use of manual labels, but also consistently improves over the self-supervised 2D baselines.
arXiv Detail & Related papers (2022-09-07T23:24:09Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - Asymmetric 3D Context Fusion for Universal Lesion Detection [55.61873234187917]
3D networks are strong in 3D context yet lack supervised pretraining.
Existing 3D context fusion operators are designed to be spatially symmetric, performing identical operations on each 2D slice like convolutions.
We propose a novel asymmetric 3D context fusion operator (A3D), which uses different weights to fuse 3D context from different 2D slices.
arXiv Detail & Related papers (2021-09-17T16:25:10Z) - Deep Hybrid Self-Prior for Full 3D Mesh Generation [57.78562932397173]
We propose to exploit a novel hybrid 2D-3D self-prior in deep neural networks to significantly improve the geometry quality.
In particular, we first generate an initial mesh using a 3D convolutional neural network with 3D self-prior, and then encode both 3D information and color information in the 2D UV atlas.
Our method recovers the 3D textured mesh model of high quality from sparse input, and outperforms the state-of-the-art methods in terms of both the geometry and texture quality.
arXiv Detail & Related papers (2021-08-18T07:44:21Z) - 3D-to-2D Distillation for Indoor Scene Parsing [78.36781565047656]
We present a new approach that enables us to leverage 3D features extracted from large-scale 3D data repository to enhance 2D features extracted from RGB images.
First, we distill 3D knowledge from a pretrained 3D network to supervise a 2D network to learn simulated 3D features from 2D features during the training.
Second, we design a two-stage dimension normalization scheme to calibrate the 2D and 3D features for better integration.
Third, we design a semantic-aware adversarial training model to extend our framework for training with unpaired 3D data.
arXiv Detail & Related papers (2021-04-06T02:22:24Z) - Bidirectional Projection Network for Cross Dimension Scene Understanding [69.29443390126805]
We present a emphbidirectional projection network (BPNet) for joint 2D and 3D reasoning in an end-to-end manner.
Via the emphBPM, complementary 2D and 3D information can interact with each other in multiple architectural levels.
Our emphBPNet achieves top performance on the ScanNetV2 benchmark for both 2D and 3D semantic segmentation.
arXiv Detail & Related papers (2021-03-26T08:31:39Z) - Pix2Vox++: Multi-scale Context-aware 3D Object Reconstruction from
Single and Multiple Images [56.652027072552606]
We propose a novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++.
By using a well-designed encoder-decoder, it generates a coarse 3D volume from each input image.
A multi-scale context-aware fusion module is then introduced to adaptively select high-quality reconstructions for different parts from all coarse 3D volumes to obtain a fused 3D volume.
arXiv Detail & Related papers (2020-06-22T13:48:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.