Learning to Accelerate Decomposition for Multi-Directional 3D Printing
- URL: http://arxiv.org/abs/2004.03450v3
- Date: Sat, 18 Jul 2020 04:50:55 GMT
- Title: Learning to Accelerate Decomposition for Multi-Directional 3D Printing
- Authors: Chenming Wu, Yong-Jin Liu, Charlie C.L. Wang
- Abstract summary: Multi-directional 3D printing has the capability of decreasing or eliminating the need for support structures.
Recent work proposed a beam-guided search algorithm to find an optimized sequence of plane-clipping.
We propose a learning framework that can accelerate the beam-guided search by using a smaller number of the original beam width.
- Score: 31.658049974100088
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-directional 3D printing has the capability of decreasing or eliminating
the need for support structures. Recent work proposed a beam-guided search
algorithm to find an optimized sequence of plane-clipping, which gives volume
decomposition of a given 3D model. Different printing directions are employed
in different regions to fabricate a model with tremendously less support (or
even no support in many cases).To obtain optimized decomposition, a large beam
width needs to be used in the search algorithm, leading to a very
time-consuming computation. In this paper, we propose a learning framework that
can accelerate the beam-guided search by using a smaller number of the original
beam width to obtain results with similar quality. Specifically, we use the
results of beam-guided search with large beam width to train a scoring function
for candidate clipping planes based on six newly proposed feature metrics. With
the help of these feature metrics, both the current and the sequence-dependent
information are captured by the neural network to score candidates of clipping.
As a result, we can achieve around 3x computational speed. We test and
demonstrate our accelerated decomposition on a large dataset of models for 3D
printing.
Related papers
- fVDB: A Deep-Learning Framework for Sparse, Large-Scale, and High-Performance Spatial Intelligence [50.417261057533786]
fVDB is a novel framework for deep learning on large-scale 3D data.
Our framework is fully integrated with PyTorch enabling interoperability with existing pipelines.
arXiv Detail & Related papers (2024-07-01T20:20:33Z) - Efficient and Distributed Large-Scale 3D Map Registration using Tomographic Features [10.740403545402508]
A robust, resource-efficient, distributed, and minimally parameterized 3D map matching and merging algorithm is proposed.
The suggested algorithm utilizes tomographic features from 2D projections of horizontal cross-sections of gravity-aligned local maps, and matches these projection slices at all possible height differences.
arXiv Detail & Related papers (2024-06-27T18:03:06Z) - N-BVH: Neural ray queries with bounding volume hierarchies [51.430495562430565]
In 3D computer graphics, the bulk of a scene's memory usage is due to polygons and textures.
We devise N-BVH, a neural compression architecture designed to answer arbitrary ray queries in 3D.
Our method provides faithful approximations of visibility, depth, and appearance attributes.
arXiv Detail & Related papers (2024-05-25T13:54:34Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - GPU optimization of the 3D Scale-invariant Feature Transform Algorithm
and a Novel BRIEF-inspired 3D Fast Descriptor [5.1537294207900715]
This work details a highly efficient implementation of the 3D scale-invariant feature transform (SIFT) algorithm, for the purpose of machine learning from large sets of medical image data.
The primary operations of the 3D SIFT code are implemented on a graphics processing unit (GPU), including convolution, sub-sampling, and 4D peak detection from scale-space pyramids.
The performance improvements are quantified in keypoint detection and image-to-image matching experiments, using 3D MRI human brain volumes of different people.
arXiv Detail & Related papers (2021-12-19T20:56:40Z) - Sketching as a Tool for Understanding and Accelerating Self-attention
for Long Sequences [52.6022911513076]
Transformer-based models are not efficient in processing long sequences due to the quadratic space and time complexity of the self-attention modules.
We propose Linformer and Informer to reduce the quadratic complexity to linear (modulo logarithmic factors) via low-dimensional projection and row selection.
Based on the theoretical analysis, we propose Skeinformer to accelerate self-attention and further improve the accuracy of matrix approximation to self-attention.
arXiv Detail & Related papers (2021-12-10T06:58:05Z) - Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data
via Differentiable Cross-Approximation [53.95297550117153]
We propose an end-to-end trainable framework that processes large-scale visual data tensors by looking emphat a fraction of their entries only.
The proposed approach is particularly useful for large-scale multidimensional grid data, and for tasks that require context over a large receptive field.
arXiv Detail & Related papers (2021-05-29T08:39:57Z) - Light3DPose: Real-time Multi-Person 3D PoseEstimation from Multiple
Views [5.510992382274774]
We present an approach to perform 3D pose estimation of multiple people from a few calibrated camera views.
Our architecture aggregates feature-maps from a 2D pose estimator backbone into a comprehensive representation of the 3D scene.
The proposed method is inherently efficient: as a pure bottom-up approach, it is computationally independent of the number of people in the scene.
arXiv Detail & Related papers (2020-04-06T14:12:19Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.