3D Part Assembly Generation with Instance Encoded Transformer
- URL: http://arxiv.org/abs/2207.01779v1
- Date: Tue, 5 Jul 2022 02:40:57 GMT
- Title: 3D Part Assembly Generation with Instance Encoded Transformer
- Authors: Rufeng Zhang, Tao Kong, Weihao Wang, Xuan Han and Mingyu You
- Abstract summary: We propose a multi-layer transformer-based framework that involves geometric and relational reasoning between parts to update the part poses iteratively.
We extend our framework to a new task called in-process part assembly.
Our method achieves far more than 10% improvements over the current state-of-the-art in multiple metrics on the public PartNet dataset.
- Score: 22.330218525999857
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: It is desirable to enable robots capable of automatic assembly. Structural
understanding of object parts plays a crucial role in this task yet remains
relatively unexplored. In this paper, we focus on the setting of furniture
assembly from a complete set of part geometries, which is essentially a 6-DoF
part pose estimation problem. We propose a multi-layer transformer-based
framework that involves geometric and relational reasoning between parts to
update the part poses iteratively. We carefully design a unique instance
encoding to solve the ambiguity between geometrically-similar parts so that all
parts can be distinguished. In addition to assembling from scratch, we extend
our framework to a new task called in-process part assembly. Analogous to
furniture maintenance, it requires robots to continue with unfinished products
and assemble the remaining parts into appropriate positions. Our method
achieves far more than 10% improvements over the current state-of-the-art in
multiple metrics on the public PartNet dataset. Extensive experiments and
quantitative comparisons demonstrate the effectiveness of the proposed
framework.
Related papers
- 3D Geometric Shape Assembly via Efficient Point Cloud Matching [59.241448711254485]
We introduce Proxy Match Transform (PMT), an approximate high-order feature transform layer that enables reliable matching between mating surfaces of parts.
Building upon PMT, we introduce a new framework, dubbed Proxy Match TransformeR (PMTR), for the geometric assembly task.
We evaluate the proposed PMTR on the large-scale 3D geometric shape assembly benchmark dataset of Breaking Bad.
arXiv Detail & Related papers (2024-07-15T08:50:02Z) - Scalable Geometric Fracture Assembly via Co-creation Space among
Assemblers [24.89380678499307]
We develop a scalable framework for geometric fracture assembly without relying on semantic information.
We introduce a novel loss function, i.e., the geometric-based collision loss, to address collision issues during the fracture assembly process.
Our framework exhibits better performance on both PartNet and Breaking Bad datasets compared to existing state-of-the-art frameworks.
arXiv Detail & Related papers (2023-12-19T17:13:51Z) - Multi-level Reasoning for Robotic Assembly: From Sequence Inference to
Contact Selection [74.40109927350856]
We present the Part Assembly Sequence Transformer (PAST) to infer assembly sequences from a target blueprint.
We then use a motion planner and optimization to generate part movements and contacts.
Experimental results show that our approach generalizes better than prior methods.
arXiv Detail & Related papers (2023-12-17T00:47:13Z) - Leveraging SE(3) Equivariance for Learning 3D Geometric Shape Assembly [7.4109730384078025]
Shape assembly aims to reassemble parts (or fragments) into a complete object.
Shape pose disentanglement of part representations is beneficial to geometric shape assembly.
We propose to leverage SE(3) equivariance for such shape pose disentanglement.
arXiv Detail & Related papers (2023-09-13T09:00:45Z) - Object Scene Representation Transformer [56.40544849442227]
We introduce Object Scene Representation Transformer (OSRT), a 3D-centric model in which individual object representations naturally emerge through novel view synthesis.
OSRT scales to significantly more complex scenes with larger diversity of objects and backgrounds than existing methods.
It is multiple orders of magnitude faster at compositional rendering thanks to its light field parametrization and the novel Slot Mixer decoder.
arXiv Detail & Related papers (2022-06-14T15:40:47Z) - 3D Compositional Zero-shot Learning with DeCompositional Consensus [102.7571947144639]
We argue that part knowledge should be composable beyond the observed object classes.
We present 3D Compositional Zero-shot Learning as a problem of part generalization from seen to unseen object classes.
arXiv Detail & Related papers (2021-11-29T16:34:53Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - RGL-NET: A Recurrent Graph Learning framework for Progressive Part
Assembly [30.143946636770025]
We tackle the problem of developing a generalized framework for assembly robust to structural variants.
Our network can learn more plausible predictions of shape structure by accounting for priorly assembled parts.
Our resulting latent space facilitates exciting applications such as shape recovery from the point-cloud components.
arXiv Detail & Related papers (2021-07-27T14:47:43Z) - Generative 3D Part Assembly via Dynamic Graph Learning [34.108515032411695]
Part assembly is a challenging yet crucial task in 3D computer vision and robotics.
We propose an assembly-oriented dynamic graph learning framework that leverages an iterative graph neural network as a backbone.
arXiv Detail & Related papers (2020-06-14T04:26:42Z) - Learning Unsupervised Hierarchical Part Decomposition of 3D Objects from
a Single RGB Image [102.44347847154867]
We propose a novel formulation that allows to jointly recover the geometry of a 3D object as a set of primitives.
Our model recovers the higher level structural decomposition of various objects in the form of a binary tree of primitives.
Our experiments on the ShapeNet and D-FAUST datasets demonstrate that considering the organization of parts indeed facilitates reasoning about 3D geometry.
arXiv Detail & Related papers (2020-04-02T17:58:05Z) - Learning 3D Part Assembly from a Single Image [20.175502864488493]
We introduce a novel problem, single-image-guided 3D part assembly, along with a learningbased solution.
We study this problem in the setting of furniture assembly from a given complete set of parts and a single image depicting the entire assembled object.
arXiv Detail & Related papers (2020-03-21T21:19:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.