Invertible Residual Network with Regularization for Effective Medical
Image Segmentation
- URL: http://arxiv.org/abs/2103.09042v1
- Date: Tue, 16 Mar 2021 13:19:59 GMT
- Title: Invertible Residual Network with Regularization for Effective Medical
Image Segmentation
- Authors: Kashu Yamazaki, Vidhiwar Singh Rathour, T.Hoang Ngan Le
- Abstract summary: Invertible neural networks have been applied to significantly reduce activation memory footprint when training neural networks with backpropagation.
We propose two versions of the invertible Residual Network, namely Partially Invertible Residual Network (Partially-InvRes) and Fully Invertible Residual Network (Fully-InvRes)
Our results indicate that by using partially/fully invertible networks as the central workhorse in volumetric segmentation, we not only reduce memory overhead but also achieve compatible segmentation performance compared against the non-invertible 3D Unet.
- Score: 2.76240219662896
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep Convolutional Neural Networks (CNNs) i.e. Residual Networks (ResNets)
have been used successfully for many computer vision tasks, but are difficult
to scale to 3D volumetric medical data. Memory is increasingly often the
bottleneck when training 3D Convolutional Neural Networks (CNNs). Recently,
invertible neural networks have been applied to significantly reduce activation
memory footprint when training neural networks with backpropagation thanks to
the invertible functions that allow retrieving its input from its output
without storing intermediate activations in memory to perform the
backpropagation.
Among many successful network architectures, 3D Unet has been established as
a standard architecture for volumetric medical segmentation. Thus, we choose 3D
Unet as a baseline for a non-invertible network and we then extend it with the
invertible residual network. In this paper, we proposed two versions of the
invertible Residual Network, namely Partially Invertible Residual Network
(Partially-InvRes) and Fully Invertible Residual Network (Fully-InvRes). In
Partially-InvRes, the invertible residual layer is defined by a technique
called additive coupling whereas in Fully-InvRes, both invertible upsampling
and downsampling operations are learned based on squeezing (known as pixel
shuffle). Furthermore, to avoid the overfitting problem because of less
training data, a variational auto-encoder (VAE) branch is added to reconstruct
the input volumetric data itself. Our results indicate that by using
partially/fully invertible networks as the central workhorse in volumetric
segmentation, we not only reduce memory overhead but also achieve compatible
segmentation performance compared against the non-invertible 3D Unet. We have
demonstrated the proposed networks on various volumetric datasets such as iSeg
2019 and BraTS 2020.
Related papers
- Fully invertible hyperbolic neural networks for segmenting large-scale surface and sub-surface data [4.1579007112499315]
This paper focuses on a fully invertible network based on the telegraph equation.
We address the explosion of convolutional kernels by combining fully invertible networks with layers that contain the convolutional kernels in a compressed form directly.
Examples in hyperspectral land-use classification, airborne geophysical surveying, and seismic imaging illustrate that we can input large data volumes in one chunk and do not need to work on small patches.
arXiv Detail & Related papers (2024-06-30T05:35:12Z) - Leveraging SO(3)-steerable convolutions for pose-robust semantic segmentation in 3D medical data [2.207533492015563]
We present a new family of segmentation networks that use equivariant voxel convolutions based on spherical harmonics.
These networks are robust to data poses not seen during training, and do not require rotation-based data augmentation during training.
We demonstrate improved segmentation performance in MRI brain tumor and healthy brain structure segmentation tasks.
arXiv Detail & Related papers (2023-03-01T09:27:08Z) - SVNet: Where SO(3) Equivariance Meets Binarization on Point Cloud
Representation [65.4396959244269]
The paper tackles the challenge by designing a general framework to construct 3D learning architectures.
The proposed approach can be applied to general backbones like PointNet and DGCNN.
Experiments on ModelNet40, ShapeNet, and the real-world dataset ScanObjectNN, demonstrated that the method achieves a great trade-off between efficiency, rotation, and accuracy.
arXiv Detail & Related papers (2022-09-13T12:12:19Z) - GLEAM: Greedy Learning for Large-Scale Accelerated MRI Reconstruction [50.248694764703714]
Unrolled neural networks have recently achieved state-of-the-art accelerated MRI reconstruction.
These networks unroll iterative optimization algorithms by alternating between physics-based consistency and neural-network based regularization.
We propose Greedy LEarning for Accelerated MRI reconstruction, an efficient training strategy for high-dimensional imaging settings.
arXiv Detail & Related papers (2022-07-18T06:01:29Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - Nonlinear Tensor Ring Network [39.89070144585793]
State-of-the-art deep neural networks (DNNs) have been widely applied for various real-world applications, and achieved significant performance for cognitive problems.
By converting redundant models into compact ones, compression technique appears to be a practical solution to reducing the storage and memory consumption.
In this paper, we develop a nonlinear tensor ring network (NTRN) in which both fullyconnected and convolutional layers are compressed.
arXiv Detail & Related papers (2021-11-12T02:02:55Z) - A New Backbone for Hyperspectral Image Reconstruction [90.48427561874402]
3D hyperspectral image (HSI) reconstruction refers to inverse process of snapshot compressive imaging.
Proposal is for a Spatial/Spectral Invariant Residual U-Net, namely SSI-ResU-Net.
We show that SSI-ResU-Net achieves competing performance with over 77.3% reduction in terms of floating-point operations.
arXiv Detail & Related papers (2021-08-17T16:20:51Z) - CoTr: Efficiently Bridging CNN and Transformer for 3D Medical Image
Segmentation [95.51455777713092]
Convolutional neural networks (CNNs) have been the de facto standard for nowadays 3D medical image segmentation.
We propose a novel framework that efficiently bridges a bf Convolutional neural network and a bf Transformer bf (CoTr) for accurate 3D medical image segmentation.
arXiv Detail & Related papers (2021-03-04T13:34:22Z) - RANP: Resource Aware Neuron Pruning at Initialization for 3D CNNs [32.054160078692036]
We introduce a Resource Aware Neuron Pruning (RANP) algorithm that prunes 3D CNNs to high sparsity levels.
Our algorithm leads to roughly 50%-95% reduction in FLOPs and 35%-80% reduction in memory with negligible loss in accuracy compared to the unpruned networks.
arXiv Detail & Related papers (2021-02-09T04:35:29Z) - Efficient Integer-Arithmetic-Only Convolutional Neural Networks [87.01739569518513]
We replace conventional ReLU with Bounded ReLU and find that the decline is due to activation quantization.
Our integer networks achieve equivalent performance as the corresponding FPN networks, but have only 1/4 memory cost and run 2x faster on modern GPU.
arXiv Detail & Related papers (2020-06-21T08:23:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.