Learning to Accelerate Decomposition for Multi-Directional 3D Printing
- URL: http://arxiv.org/abs/2004.03450v3
- Date: Sat, 18 Jul 2020 04:50:55 GMT
- Title: Learning to Accelerate Decomposition for Multi-Directional 3D Printing
- Authors: Chenming Wu, Yong-Jin Liu, Charlie C.L. Wang
- Abstract summary: Multi-directional 3D printing has the capability of decreasing or eliminating the need for support structures.
Recent work proposed a beam-guided search algorithm to find an optimized sequence of plane-clipping.
We propose a learning framework that can accelerate the beam-guided search by using a smaller number of the original beam width.
- Score: 31.658049974100088
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-directional 3D printing has the capability of decreasing or eliminating
the need for support structures. Recent work proposed a beam-guided search
algorithm to find an optimized sequence of plane-clipping, which gives volume
decomposition of a given 3D model. Different printing directions are employed
in different regions to fabricate a model with tremendously less support (or
even no support in many cases).To obtain optimized decomposition, a large beam
width needs to be used in the search algorithm, leading to a very
time-consuming computation. In this paper, we propose a learning framework that
can accelerate the beam-guided search by using a smaller number of the original
beam width to obtain results with similar quality. Specifically, we use the
results of beam-guided search with large beam width to train a scoring function
for candidate clipping planes based on six newly proposed feature metrics. With
the help of these feature metrics, both the current and the sequence-dependent
information are captured by the neural network to score candidates of clipping.
As a result, we can achieve around 3x computational speed. We test and
demonstrate our accelerated decomposition on a large dataset of models for 3D
printing.
Related papers
- Pseudo Depth Meets Gaussian: A Feed-forward RGB SLAM Baseline [64.42938561167402]
We propose an online 3D reconstruction method using 3D Gaussian-based SLAM, combined with a feed-forward recurrent prediction module.<n>This approach replaces slow test-time optimization with fast network inference, significantly improving tracking speed.<n>Our method achieves performance on par with the state-of-the-art SplaTAM, while reducing tracking time by more than 90%.
arXiv Detail & Related papers (2025-08-06T16:16:58Z) - PointVDP: Learning View-Dependent Projection by Fireworks Rays for 3D Point Cloud Segmentation [66.00721801098574]
We propose view-dependent projection (VDP) to facilitate point cloud segmentation.<n>VDP generates data-driven projections from 3D point distributions.<n>We construct color regularization to optimize the framework.
arXiv Detail & Related papers (2025-07-09T07:44:00Z) - Dynamic-Width Speculative Beam Decoding for Efficient LLM Inference [35.730941605490194]
Large language models (LLMs) have shown outstanding performance across numerous real-world tasks.
Speculative decoding has emerged as a promising solution, leveraging a smaller auxiliary model to draft future tokens.
This paper explores the novel integration of speculative decoding with beam sampling.
arXiv Detail & Related papers (2024-09-25T02:20:42Z) - fVDB: A Deep-Learning Framework for Sparse, Large-Scale, and High-Performance Spatial Intelligence [50.417261057533786]
fVDB is a novel framework for deep learning on large-scale 3D data.
Our framework is fully integrated with PyTorch enabling interoperability with existing pipelines.
arXiv Detail & Related papers (2024-07-01T20:20:33Z) - Efficient and Distributed Large-Scale 3D Map Registration using Tomographic Features [10.740403545402508]
A robust, resource-efficient, distributed, and minimally parameterized 3D map matching and merging algorithm is proposed.
The suggested algorithm utilizes tomographic features from 2D projections of horizontal cross-sections of gravity-aligned local maps, and matches these projection slices at all possible height differences.
arXiv Detail & Related papers (2024-06-27T18:03:06Z) - N-BVH: Neural ray queries with bounding volume hierarchies [51.430495562430565]
In 3D computer graphics, the bulk of a scene's memory usage is due to polygons and textures.
We devise N-BVH, a neural compression architecture designed to answer arbitrary ray queries in 3D.
Our method provides faithful approximations of visibility, depth, and appearance attributes.
arXiv Detail & Related papers (2024-05-25T13:54:34Z) - *: Improving the 3D detector by introducing Voxel2Pillar feature encoding and extracting multi-scale features [9.15169530632709]
Current 3D detectors commonly use feature pyramid networks to obtain large-scale features.
Since pillar-based schemes require much less than voxel computation schemes, they are more suitable for constructing realtime 3D detectors.
We propose the Voxel2Pillar feature encoding, which uses a sparse convolution to construct pillars with richer point cloud features.
arXiv Detail & Related papers (2024-05-16T06:05:08Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - Sketching as a Tool for Understanding and Accelerating Self-attention
for Long Sequences [52.6022911513076]
Transformer-based models are not efficient in processing long sequences due to the quadratic space and time complexity of the self-attention modules.
We propose Linformer and Informer to reduce the quadratic complexity to linear (modulo logarithmic factors) via low-dimensional projection and row selection.
Based on the theoretical analysis, we propose Skeinformer to accelerate self-attention and further improve the accuracy of matrix approximation to self-attention.
arXiv Detail & Related papers (2021-12-10T06:58:05Z) - Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data
via Differentiable Cross-Approximation [53.95297550117153]
We propose an end-to-end trainable framework that processes large-scale visual data tensors by looking emphat a fraction of their entries only.
The proposed approach is particularly useful for large-scale multidimensional grid data, and for tasks that require context over a large receptive field.
arXiv Detail & Related papers (2021-05-29T08:39:57Z) - Light3DPose: Real-time Multi-Person 3D PoseEstimation from Multiple
Views [5.510992382274774]
We present an approach to perform 3D pose estimation of multiple people from a few calibrated camera views.
Our architecture aggregates feature-maps from a 2D pose estimator backbone into a comprehensive representation of the 3D scene.
The proposed method is inherently efficient: as a pure bottom-up approach, it is computationally independent of the number of people in the scene.
arXiv Detail & Related papers (2020-04-06T14:12:19Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.