HyperCube: Implicit Field Representations of Voxelized 3D Models
- URL: http://arxiv.org/abs/2110.05770v1
- Date: Tue, 12 Oct 2021 06:56:48 GMT
- Title: HyperCube: Implicit Field Representations of Voxelized 3D Models
- Authors: Magdalena Proszewska, Marcin Mazur, Tomasz Trzci\'nski, Przemys{\l}aw
Spurek
- Abstract summary: We introduce a new HyperCube architecture that enables direct processing of 3D voxels.
Instead of processing individual 3D samples from within a voxel, our approach allows to input the entire voxel represented with its convex hull coordinates.
- Score: 18.868266675878996
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently introduced implicit field representations offer an effective way of
generating 3D object shapes. They leverage implicit decoder trained to take a
3D point coordinate concatenated with a shape encoding and to output a value
which indicates whether the point is outside the shape or not. Although this
approach enables efficient rendering of visually plausible objects, it has two
significant limitations. First, it is based on a single neural network
dedicated for all objects from a training set which results in a cumbersome
training procedure and its application in real life. More importantly, the
implicit decoder takes only points sampled within voxels (and not the entire
voxels) which yields problems at the classification boundaries and results in
empty spaces within the rendered mesh.
To solve the above limitations, we introduce a new HyperCube architecture
based on interval arithmetic network, that enables direct processing of 3D
voxels, trained using a hypernetwork paradigm to enforce model convergence.
Instead of processing individual 3D samples from within a voxel, our approach
allows to input the entire voxel (3D cube) represented with its convex hull
coordinates, while the target network constructed by a hypernet assigns it to
an inside or outside category. As a result our HyperCube model outperforms the
competing approaches both in terms of training and inference efficiency, as
well as the final mesh quality.
Related papers
- MeshXL: Neural Coordinate Field for Generative 3D Foundation Models [51.1972329762843]
We present a family of generative pre-trained auto-regressive models, which addresses the process of 3D mesh generation with modern large language model approaches.
MeshXL is able to generate high-quality 3D meshes, and can also serve as foundation models for various down-stream applications.
arXiv Detail & Related papers (2024-05-31T14:35:35Z) - Oriented-grid Encoder for 3D Implicit Representations [10.02138130221506]
This paper is the first to exploit 3D characteristics in 3D geometric encoders explicitly.
Our method gets state-of-the-art results when compared to the prior techniques.
arXiv Detail & Related papers (2024-02-09T19:28:13Z) - Model2Scene: Learning 3D Scene Representation via Contrastive
Language-CAD Models Pre-training [105.3421541518582]
Current successful methods of 3D scene perception rely on the large-scale annotated point cloud.
We propose Model2Scene, a novel paradigm that learns free 3D scene representation from Computer-Aided Design (CAD) models and languages.
Model2Scene yields impressive label-free 3D object salient detection with an average mAP of 46.08% and 55.49% on the ScanNet and S3DIS datasets, respectively.
arXiv Detail & Related papers (2023-09-29T03:51:26Z) - Deep Marching Tetrahedra: a Hybrid Representation for High-Resolution 3D
Shape Synthesis [90.26556260531707]
DMTet is a conditional generative model that can synthesize high-resolution 3D shapes using simple user guides such as coarse voxels.
Unlike deep 3D generative models that directly generate explicit representations such as meshes, our model can synthesize shapes with arbitrary topology.
arXiv Detail & Related papers (2021-11-08T05:29:35Z) - HVPR: Hybrid Voxel-Point Representation for Single-stage 3D Object
Detection [39.64891219500416]
3D object detection methods exploit either voxel-based or point-based features to represent 3D objects in a scene.
We introduce in this paper a novel single-stage 3D detection method having the merit of both voxel-based and point-based features.
arXiv Detail & Related papers (2021-04-02T06:34:49Z) - A Fast Hybrid Cascade Network for Voxel-based 3D Object Classification [10.019858113123822]
We propose a hybrid cascade architecture for voxel-based 3D object classification.
Both accuracy and speed can be balanced in our proposed method.
arXiv Detail & Related papers (2020-11-09T15:58:33Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z) - HyperFlow: Representing 3D Objects as Surfaces [19.980044265074298]
We present a novel generative model that leverages hypernetworks to create continuous 3D object representations in a form of lightweight surfaces (meshes) directly out of point clouds.
We obtain continuous mesh-based object representations that yield better qualitative results than competing approaches.
arXiv Detail & Related papers (2020-06-15T19:18:02Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z) - Convolutional Occupancy Networks [88.48287716452002]
We propose Convolutional Occupancy Networks, a more flexible implicit representation for detailed reconstruction of objects and 3D scenes.
By combining convolutional encoders with implicit occupancy decoders, our model incorporates inductive biases, enabling structured reasoning in 3D space.
We empirically find that our method enables the fine-grained implicit 3D reconstruction of single objects, scales to large indoor scenes, and generalizes well from synthetic to real data.
arXiv Detail & Related papers (2020-03-10T10:17:07Z) - Hypernetwork approach to generating point clouds [18.67883065951206]
We build a hyper network that returns weights of a particular neural network trained to map points into a 3D shape.
A particular 3D shape can be generated using point-by-point sampling from the assumed prior distribution.
Since the hyper network is based on an auto-encoder architecture trained to reconstruct realistic 3D shapes, the target network weights can be considered a parametrization of the surface of a 3D shape.
arXiv Detail & Related papers (2020-02-10T11:09:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.