State of the Art in Dense Monocular Non-Rigid 3D Reconstruction
- URL: http://arxiv.org/abs/2210.15664v2
- Date: Fri, 24 Mar 2023 18:45:56 GMT
- Title: State of the Art in Dense Monocular Non-Rigid 3D Reconstruction
- Authors: Edith Tretschk, Navami Kairanda, Mallikarjun B R, Rishabh Dabral, Adam
Kortylewski, Bernhard Egger, Marc Habermann, Pascal Fua, Christian Theobalt,
Vladislav Golyanik
- Abstract summary: 3D reconstruction of deformable (or non-rigid) scenes from a set of monocular 2D image observations is a long-standing and actively researched area of computer vision and graphics.
This survey focuses on state-of-the-art methods for dense non-rigid 3D reconstruction of various deformable objects and composite scenes from monocular videos or sets of monocular views.
- Score: 100.9586977875698
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D reconstruction of deformable (or non-rigid) scenes from a set of monocular
2D image observations is a long-standing and actively researched area of
computer vision and graphics. It is an ill-posed inverse problem, since --
without additional prior assumptions -- it permits infinitely many solutions
leading to accurate projection to the input 2D images. Non-rigid reconstruction
is a foundational building block for downstream applications like robotics,
AR/VR, or visual content creation. The key advantage of using monocular cameras
is their omnipresence and availability to the end users as well as their ease
of use compared to more sophisticated camera set-ups such as stereo or
multi-view systems. This survey focuses on state-of-the-art methods for dense
non-rigid 3D reconstruction of various deformable objects and composite scenes
from monocular videos or sets of monocular views. It reviews the fundamentals
of 3D reconstruction and deformation modeling from 2D image observations. We
then start from general methods -- that handle arbitrary scenes and make only a
few prior assumptions -- and proceed towards techniques making stronger
assumptions about the observed objects and types of deformations (e.g. human
faces, bodies, hands, and animals). A significant part of this STAR is also
devoted to classification and a high-level comparison of the methods, as well
as an overview of the datasets for training and evaluation of the discussed
techniques. We conclude by discussing open challenges in the field and the
social aspects associated with the usage of the reviewed methods.
Related papers
- 3D Scene Geometry Estimation from 360$^\circ$ Imagery: A Survey [1.3654846342364308]
This paper provides a comprehensive survey on pioneer and state-of-the-art 3D scene geometry estimation methodologies.
We first revisit the basic concepts of the spherical camera model, and review the most common acquisition technologies and representation formats.
We then survey monocular layout and depth inference approaches, highlighting the recent advances in learning-based solutions suited for spherical data.
arXiv Detail & Related papers (2024-01-17T14:57:27Z) - DUSt3R: Geometric 3D Vision Made Easy [8.471330244002564]
We introduce DUSt3R, a novel paradigm for Dense and Unconstrained Stereo 3D Reconstruction of arbitrary image collections.
We show that this formulation smoothly unifies the monocular and binocular reconstruction cases.
Our formulation directly provides a 3D model of the scene as well as depth information, but interestingly, we can seamlessly recover from it, pixel matches, relative and absolute camera.
arXiv Detail & Related papers (2023-12-21T18:52:14Z) - Decaf: Monocular Deformation Capture for Face and Hand Interactions [77.75726740605748]
This paper introduces the first method that allows tracking human hands interacting with human faces in 3D from single monocular RGB videos.
We model hands as articulated objects inducing non-rigid face deformations during an active interaction.
Our method relies on a new hand-face motion and interaction capture dataset with realistic face deformations acquired with a markerless multi-view camera system.
arXiv Detail & Related papers (2023-09-28T17:59:51Z) - Multiview Compressive Coding for 3D Reconstruction [77.95706553743626]
We introduce a simple framework that operates on 3D points of single objects or whole scenes.
Our model, Multiview Compressive Coding, learns to compress the input appearance and geometry to predict the 3D structure.
arXiv Detail & Related papers (2023-01-19T18:59:52Z) - Few-View Object Reconstruction with Unknown Categories and Camera Poses [80.0820650171476]
This work explores reconstructing general real-world objects from a few images without known camera poses or object categories.
The crux of our work is solving two fundamental 3D vision problems -- shape reconstruction and pose estimation.
Our method FORGE predicts 3D features from each view and leverages them in conjunction with the input images to establish cross-view correspondence.
arXiv Detail & Related papers (2022-12-08T18:59:02Z) - 3D Magic Mirror: Clothing Reconstruction from a Single Image via a
Causal Perspective [96.65476492200648]
This research aims to study a self-supervised 3D clothing reconstruction method.
It recovers the geometry shape, and texture of human clothing from a single 2D image.
arXiv Detail & Related papers (2022-04-27T17:46:55Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.