Deep Direct Volume Rendering: Learning Visual Feature Mappings From
Exemplary Images
- URL: http://arxiv.org/abs/2106.05429v1
- Date: Wed, 9 Jun 2021 23:03:00 GMT
- Title: Deep Direct Volume Rendering: Learning Visual Feature Mappings From
Exemplary Images
- Authors: Jakob Weiss, Nassir Navab
- Abstract summary: We introduce Deep Direct Volume Rendering (DeepDVR), a generalization of Direct Volume Rendering (DVR) that allows for the integration of deep neural networks into the DVR algorithm.
We conceptualize the rendering in a latent color space, thus enabling the use of deep architectures to learn implicit mappings for feature extraction and classification.
Our generalization serves to derive novel volume rendering architectures that can be trained end-to-end directly from examples in image space.
- Score: 57.253447453301796
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Volume Rendering is an important technique for visualizing three-dimensional
scalar data grids and is commonly employed for scientific and medical image
data. Direct Volume Rendering (DVR) is a well established and efficient
rendering algorithm for volumetric data. Neural rendering uses deep neural
networks to solve inverse rendering tasks and applies techniques similar to
DVR. However, it has not been demonstrated successfully for the rendering of
scientific volume data.
In this work, we introduce Deep Direct Volume Rendering (DeepDVR), a
generalization of DVR that allows for the integration of deep neural networks
into the DVR algorithm. We conceptualize the rendering in a latent color space,
thus enabling the use of deep architectures to learn implicit mappings for
feature extraction and classification, replacing explicit feature design and
hand-crafted transfer functions. Our generalization serves to derive novel
volume rendering architectures that can be trained end-to-end directly from
examples in image space, obviating the need to manually define and fine-tune
multidimensional transfer functions while providing superior classification
strength. We further introduce a novel stepsize annealing scheme to accelerate
the training of DeepDVR models and validate its effectiveness in a set of
experiments. We validate our architectures on two example use cases: (1)
learning an optimized rendering from manually adjusted reference images for a
single volume and (2) learning advanced visualization concepts like shading and
semantic colorization that generalize to unseen volume data.
We find that deep volume rendering architectures with explicit modeling of
the DVR pipeline effectively enable end-to-end learning of scientific volume
rendering tasks from target images.
Related papers
- TK-Planes: Tiered K-Planes with High Dimensional Feature Vectors for Dynamic UAV-based Scenes [58.180556221044235]
We present a new approach to bridge the domain gap between synthetic and real-world data for unmanned aerial vehicle (UAV)-based perception.
Our formulation is designed for dynamic scenes, consisting of small moving objects or human actions.
We evaluate its performance on challenging datasets, including Okutama Action and UG2.
arXiv Detail & Related papers (2024-05-04T21:55:33Z) - DNS SLAM: Dense Neural Semantic-Informed SLAM [92.39687553022605]
DNS SLAM is a novel neural RGB-D semantic SLAM approach featuring a hybrid representation.
Our method integrates multi-view geometry constraints with image-based feature extraction to improve appearance details.
Our experimental results achieve state-of-the-art performance on both synthetic data and real-world data tracking.
arXiv Detail & Related papers (2023-11-30T21:34:44Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - FoVolNet: Fast Volume Rendering using Foveated Deep Neural Networks [33.489890950757975]
FoVolNet is a method to significantly increase the performance of volume data visualization.
We develop a cost-effective foveated rendering pipeline that sparsely samples a volume around a focal point and reconstructs the full-frame using a deep neural network.
arXiv Detail & Related papers (2022-09-20T19:48:56Z) - 3DVNet: Multi-View Depth Prediction and Volumetric Refinement [68.68537312256144]
3DVNet is a novel multi-view stereo (MVS) depth-prediction method.
Our key idea is the use of a 3D scene-modeling network that iteratively updates a set of coarse depth predictions.
We show that our method exceeds state-of-the-art accuracy in both depth prediction and 3D reconstruction metrics.
arXiv Detail & Related papers (2021-12-01T00:52:42Z) - S2R-DepthNet: Learning a Generalizable Depth-specific Structural
Representation [63.58891781246175]
Human can infer the 3D geometry of a scene from a sketch instead of a realistic image, which indicates that the spatial structure plays a fundamental role in understanding the depth of scenes.
We are the first to explore the learning of a depth-specific structural representation, which captures the essential feature for depth estimation and ignores irrelevant style information.
Our S2R-DepthNet can be well generalized to unseen real-world data directly even though it is only trained on synthetic data.
arXiv Detail & Related papers (2021-04-02T03:55:41Z) - Physics-based Differentiable Depth Sensor Simulation [5.134435281973137]
We introduce a novel end-to-end differentiable simulation pipeline for the generation of realistic 2.5D scans.
Each module can be differentiated w.r.t sensor and scene parameters.
Our simulation greatly improves the performance of the resulting models on real scans.
arXiv Detail & Related papers (2021-03-30T17:59:43Z) - On Deep Learning Techniques to Boost Monocular Depth Estimation for
Autonomous Navigation [1.9007546108571112]
Inferring the depth of images is a fundamental inverse problem within the field of Computer Vision.
We propose a new lightweight and fast supervised CNN architecture combined with novel feature extraction models.
We also introduce an efficient surface normals module, jointly with a simple geometric 2.5D loss function, to solve SIDE problems.
arXiv Detail & Related papers (2020-10-13T18:37:38Z) - VC-Net: Deep Volume-Composition Networks for Segmentation and
Visualization of Highly Sparse and Noisy Image Data [13.805816310795256]
We present an end-to-end deep learning method, VC-Net, for robust extraction of 3D microvasculature.
The core novelty is to automatically leverage the volume visualization technique (MIP) to enhance the 3D data exploration.
A multi-stream convolutional neural network is proposed to learn the 3D volume and 2D MIP features respectively and then explore their inter-dependencies in a joint volume-composition embedding space.
arXiv Detail & Related papers (2020-09-14T04:15:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.