Occupancy Planes for Single-view RGB-D Human Reconstruction
- URL: http://arxiv.org/abs/2208.02817v1
- Date: Thu, 4 Aug 2022 17:59:56 GMT
- Title: Occupancy Planes for Single-view RGB-D Human Reconstruction
- Authors: Xiaoming Zhao and Yuan-Ting Hu and Zhongzheng Ren and Alexander G.
Schwing
- Abstract summary: Single-view RGB-D human reconstruction with implicit functions is often formulated as per-point classification.
We propose the occupancy planes (OPlanes) representation, which enables to formulate single-view RGB-D human reconstruction as occupancy prediction on planes which slice through the camera's view frustum.
- Score: 120.5818162569105
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Single-view RGB-D human reconstruction with implicit functions is often
formulated as per-point classification. Specifically, a set of 3D locations
within the view-frustum of the camera are first projected independently onto
the image and a corresponding feature is subsequently extracted for each 3D
location. The feature of each 3D location is then used to classify
independently whether the corresponding 3D point is inside or outside the
observed object. This procedure leads to sub-optimal results because
correlations between predictions for neighboring locations are only taken into
account implicitly via the extracted features. For more accurate results we
propose the occupancy planes (OPlanes) representation, which enables to
formulate single-view RGB-D human reconstruction as occupancy prediction on
planes which slice through the camera's view frustum. Such a representation
provides more flexibility than voxel grids and enables to better leverage
correlations than per-point classification. On the challenging S3D data we
observe a simple classifier based on the OPlanes representation to yield
compelling results, especially in difficult situations with partial occlusions
due to other objects and partial visibility, which haven't been addressed by
prior work.
Related papers
- Normal-guided Detail-Preserving Neural Implicit Functions for High-Fidelity 3D Surface Reconstruction [6.4279213810512665]
Current methods for learning neural implicit representations from RGB or RGBD images produce 3D surfaces with missing parts and details.
This paper demonstrates that training neural representations with first-order differential properties, i.e. surface normals, leads to highly accurate 3D surface reconstruction.
arXiv Detail & Related papers (2024-06-07T11:48:47Z) - Occ$^2$Net: Robust Image Matching Based on 3D Occupancy Estimation for
Occluded Regions [14.217367037250296]
Occ$2$Net is an image matching method that models occlusion relations using 3D occupancy and infers matching points in occluded regions.
We evaluate our method on both real-world and simulated datasets and demonstrate its superior performance over state-of-the-art methods on several metrics.
arXiv Detail & Related papers (2023-08-14T13:09:41Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - Neural Correspondence Field for Object Pose Estimation [67.96767010122633]
We propose a method for estimating the 6DoF pose of a rigid object with an available 3D model from a single RGB image.
Unlike classical correspondence-based methods which predict 3D object coordinates at pixels of the input image, the proposed method predicts 3D object coordinates at 3D query points sampled in the camera frustum.
arXiv Detail & Related papers (2022-07-30T01:48:23Z) - BiCo-Net: Regress Globally, Match Locally for Robust 6D Pose Estimation [32.49091033895255]
Bi-directional Correspondence Mapping Network (BiCo-Net) generates point clouds guided by a typical pose regression.
An ensemble of redundant pose predictions from locally matching and direct pose regression further refines final pose output against noisy observations.
arXiv Detail & Related papers (2022-05-07T03:37:33Z) - What's in your hands? 3D Reconstruction of Generic Objects in Hands [49.12461675219253]
Our work aims to reconstruct hand-held objects given a single RGB image.
In contrast to prior works that typically assume known 3D templates and reduce the problem to 3D pose estimation, our work reconstructs generic hand-held object without knowing their 3D templates.
arXiv Detail & Related papers (2022-04-14T17:59:02Z) - Stochastic Modeling for Learnable Human Pose Triangulation [0.7646713951724009]
We propose a modeling framework for 3D human pose triangulation and evaluate its performance across different datasets and spatial camera arrangements.
The proposed pose triangulation model successfully generalizes to different camera arrangements and between two public datasets.
arXiv Detail & Related papers (2021-10-01T09:26:25Z) - Object-Centric Multi-View Aggregation [86.94544275235454]
We present an approach for aggregating a sparse set of views of an object in order to compute a semi-implicit 3D representation in the form of a volumetric feature grid.
Key to our approach is an object-centric canonical 3D coordinate system into which views can be lifted, without explicit camera pose estimation.
We show that computing a symmetry-aware mapping from pixels to the canonical coordinate system allows us to better propagate information to unseen regions.
arXiv Detail & Related papers (2020-07-20T17:38:31Z) - Convolutional Occupancy Networks [88.48287716452002]
We propose Convolutional Occupancy Networks, a more flexible implicit representation for detailed reconstruction of objects and 3D scenes.
By combining convolutional encoders with implicit occupancy decoders, our model incorporates inductive biases, enabling structured reasoning in 3D space.
We empirically find that our method enables the fine-grained implicit 3D reconstruction of single objects, scales to large indoor scenes, and generalizes well from synthetic to real data.
arXiv Detail & Related papers (2020-03-10T10:17:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.