Object-Compositional Neural Implicit Surfaces
- URL: http://arxiv.org/abs/2207.09686v1
- Date: Wed, 20 Jul 2022 06:38:04 GMT
- Title: Object-Compositional Neural Implicit Surfaces
- Authors: Qianyi Wu, Xian Liu, Yuedong Chen, Kejie Li, Chuanxia Zheng, Jianfei
Cai, Jianmin Zheng
- Abstract summary: The neural implicit representation has shown its effectiveness in novel view synthesis and high-quality 3D reconstruction from multi-view images.
This paper proposes a novel framework, ObjectSDF, to build an object-compositional neural implicit representation with high fidelity in 3D reconstruction and object representation.
- Score: 45.274466719163925
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The neural implicit representation has shown its effectiveness in novel view
synthesis and high-quality 3D reconstruction from multi-view images. However,
most approaches focus on holistic scene representation yet ignore individual
objects inside it, thus limiting potential downstream applications. In order to
learn object-compositional representation, a few works incorporate the 2D
semantic map as a cue in training to grasp the difference between objects. But
they neglect the strong connections between object geometry and instance
semantic information, which leads to inaccurate modeling of individual
instance. This paper proposes a novel framework, ObjectSDF, to build an
object-compositional neural implicit representation with high fidelity in 3D
reconstruction and object representation. Observing the ambiguity of
conventional volume rendering pipelines, we model the scene by combining the
Signed Distance Functions (SDF) of individual object to exert explicit surface
constraint. The key in distinguishing different instances is to revisit the
strong association between an individual object's SDF and semantic label.
Particularly, we convert the semantic information to a function of object SDF
and develop a unified and compact representation for scene and objects.
Experimental results show the superiority of ObjectSDF framework in
representing both the holistic object-compositional scene and the individual
instances. Code can be found at https://qianyiwu.github.io/objectsdf/
Related papers
- Object Pose Estimation Using Implicit Representation For Transparent Objects [0.0]
The render-and-compare method renders the object from multiple views and compares it against the given 2D image.
We show that if the object is represented as an implicit (neural) representation in the form of Neural Radiance Field (NeRF), it exhibits a more realistic rendering of the actual scene.
We evaluated our NeRF implementation of the render-and-compare method on transparent datasets and found that it surpassed the current state-of-the-art results.
arXiv Detail & Related papers (2024-10-17T11:51:12Z) - Chat-Scene: Bridging 3D Scene and Large Language Models with Object Identifiers [65.51132104404051]
We introduce the use of object identifiers and object-centric representations to interact with scenes at the object level.
Our model significantly outperforms existing methods on benchmarks including ScanRefer, Multi3DRefer, Scan2Cap, ScanQA, and SQA3D.
arXiv Detail & Related papers (2023-12-13T14:27:45Z) - Iterative Superquadric Recomposition of 3D Objects from Multiple Views [77.53142165205283]
We propose a framework, ISCO, to recompose an object using 3D superquadrics as semantic parts directly from 2D views.
Our framework iteratively adds new superquadrics wherever the reconstruction error is high.
It provides consistently more accurate 3D reconstructions, even from images in the wild.
arXiv Detail & Related papers (2023-09-05T10:21:37Z) - ROAM: Robust and Object-Aware Motion Generation Using Neural Pose
Descriptors [73.26004792375556]
This paper shows that robustness and generalisation to novel scene objects in 3D object-aware character synthesis can be achieved by training a motion model with as few as one reference object.
We leverage an implicit feature representation trained on object-only datasets, which encodes an SE(3)-equivariant descriptor field around the object.
We demonstrate substantial improvements in 3D virtual character motion and interaction quality and robustness to scenarios with unseen objects.
arXiv Detail & Related papers (2023-08-24T17:59:51Z) - ObjectSDF++: Improved Object-Compositional Neural Implicit Surfaces [40.489487738598825]
In recent years, neural implicit surface reconstruction has emerged as a popular paradigm for multi-view 3D reconstruction.
Previous work ObjectSDF introduced a nice framework of object-composition neural implicit surfaces.
We propose a new framework called ObjectSDF++ to overcome the limitations of ObjectSDF.
arXiv Detail & Related papers (2023-08-15T16:35:40Z) - Hyperbolic Contrastive Learning for Visual Representations beyond
Objects [30.618032825306187]
We focus on learning representations for objects and scenes that preserve the structure among them.
Motivated by the observation that visually similar objects are close in the representation space, we argue that the scenes and objects should instead follow a hierarchical structure.
arXiv Detail & Related papers (2022-12-01T16:58:57Z) - Neural Descriptor Fields: SE(3)-Equivariant Object Representations for
Manipulation [75.83319382105894]
We present Neural Descriptor Fields (NDFs), an object representation that encodes both points and relative poses between an object and a target.
NDFs are trained in a self-supervised fashion via a 3D auto-encoding task that does not rely on expert-labeled keypoints.
Our performance generalizes across both object instances and 6-DoF object poses, and significantly outperforms a recent baseline that relies on 2D descriptors.
arXiv Detail & Related papers (2021-12-09T18:57:15Z) - TSDF++: A Multi-Object Formulation for Dynamic Object Tracking and
Reconstruction [57.1209039399599]
We propose a map representation that allows maintaining a single volume for the entire scene and all the objects therein.
In a multiple dynamic object tracking and reconstruction scenario, our representation allows maintaining accurate reconstruction of surfaces even while they become temporarily occluded by other objects moving in their proximity.
We evaluate the proposed TSDF++ formulation on a public synthetic dataset and demonstrate its ability to preserve reconstructions of occluded surfaces when compared to the standard TSDF map representation.
arXiv Detail & Related papers (2021-05-16T16:15:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.