iWave3D: End-to-end Brain Image Compression with Trainable 3-D Wavelet
Transform
- URL: http://arxiv.org/abs/2109.08942v1
- Date: Sat, 18 Sep 2021 14:38:59 GMT
- Title: iWave3D: End-to-end Brain Image Compression with Trainable 3-D Wavelet
Transform
- Authors: Dongmei Xue, Haichuan Ma, Li Li, Dong Liu, Zhiwei Xiong
- Abstract summary: We propose a trainable 3-D wavelet transform based on the lifting scheme, in which the predict and update steps are replaced by 3-D convolutional neural networks.
Experimental results demonstrate that our method outperforms JP3D significantly by 2.012 dB in terms of average BD-PSNR.
- Score: 42.14812529290784
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rapid development of whole brain imaging technology, a large number
of brain images have been produced, which puts forward a great demand for
efficient brain image compression methods. At present, the most commonly used
compression methods are all based on 3-D wavelet transform, such as JP3D.
However, traditional 3-D wavelet transforms are designed manually with certain
assumptions on the signal, but brain images are not as ideal as assumed. What's
more, they are not directly optimized for compression task. In order to solve
these problems, we propose a trainable 3-D wavelet transform based on the
lifting scheme, in which the predict and update steps are replaced by 3-D
convolutional neural networks. Then the proposed transform is embedded into an
end-to-end compression scheme called iWave3D, which is trained with a large
amount of brain images to directly minimize the rate-distortion loss.
Experimental results demonstrate that our method outperforms JP3D significantly
by 2.012 dB in terms of average BD-PSNR.
Related papers
- Deep Diffusion Image Prior for Efficient OOD Adaptation in 3D Inverse Problems [61.85478918618346]
We propose DDIP, which generalizes the recent adaptation method of SCD by introducing a formal connection to the deep image prior.
Under this framework, we propose an efficient adaptation method dubbed D3IP, specified for 3D measurements, which accelerates DDIP by orders of magnitude.
We show that our method is capable of solving diverse 3D reconstructive tasks from the generative prior trained only with phantom images that are vastly different from the training set.
arXiv Detail & Related papers (2024-07-15T12:00:46Z) - NeRFCodec: Neural Feature Compression Meets Neural Radiance Fields for Memory-Efficient Scene Representation [22.151167286623416]
We propose an end-to-end NeRF compression framework that integrates non-linear transform, quantization, and entropy coding for memory-efficient scene representation.
We demonstrate our method outperforms existing NeRF compression methods, enabling high-quality novel view synthesis with a memory budget of 0.5 MB.
arXiv Detail & Related papers (2024-04-02T15:49:00Z) - Real-Time Radiance Fields for Single-Image Portrait View Synthesis [85.32826349697972]
We present a one-shot method to infer and render a 3D representation from a single unposed image in real-time.
Given a single RGB input, our image encoder directly predicts a canonical triplane representation of a neural radiance field for 3D-aware novel view synthesis via volume rendering.
Our method is fast (24 fps) on consumer hardware, and produces higher quality results than strong GAN-inversion baselines that require test-time optimization.
arXiv Detail & Related papers (2023-05-03T17:56:01Z) - SnakeVoxFormer: Transformer-based Single Image\\Voxel Reconstruction
with Run Length Encoding [9.691609196086015]
SnakeVoxFormer is a novel, 3D object reconstruction in voxel space from a single image using the transformer.
We show how different voxel traversing strategies affect the effect of encoding and reconstruction.
arXiv Detail & Related papers (2023-03-28T20:16:13Z) - Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models [52.529394863331326]
We propose a novel approach using two perpendicular pre-trained 2D diffusion models to solve the 3D inverse problem.
Our method is highly effective for 3D medical image reconstruction tasks, including MRI Z-axis super-resolution, compressed sensing MRI, and sparse-view CT.
arXiv Detail & Related papers (2023-03-15T08:28:06Z) - Inflating 2D Convolution Weights for Efficient Generation of 3D Medical
Images [35.849240945334]
Two problems prevent effective training of a 3D medical generative model: 3D medical images are expensive to acquire and annotate, and a large number of parameters are involved in 3D convolution.
We propose a novel GAN model called 3D Split&Shuffle-GAN.
We show that our method leads to improved 3D image generation quality with significantly fewer parameters.
arXiv Detail & Related papers (2022-08-08T06:31:00Z) - aiWave: Volumetric Image Compression with 3-D Trained Affine
Wavelet-like Transform [43.984890290691695]
Most commonly used volumetric image compression methods are based on wavelet transform, such as JP3D.
In this paper, we first design a 3-D trained wavelet-like transform to enable signal-dependent and non-separable transform.
Then, an affine wavelet basis is introduced to capture the various local correlations in different regions of volumetric images.
arXiv Detail & Related papers (2022-03-11T10:02:01Z) - Pix2Vox++: Multi-scale Context-aware 3D Object Reconstruction from
Single and Multiple Images [56.652027072552606]
We propose a novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++.
By using a well-designed encoder-decoder, it generates a coarse 3D volume from each input image.
A multi-scale context-aware fusion module is then introduced to adaptively select high-quality reconstructions for different parts from all coarse 3D volumes to obtain a fused 3D volume.
arXiv Detail & Related papers (2020-06-22T13:48:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.