Fast-Image2Point: Towards Real-Time Point Cloud Reconstruction of a
Single Image using 3D Supervision
- URL: http://arxiv.org/abs/2209.10029v1
- Date: Tue, 20 Sep 2022 22:39:14 GMT
- Title: Fast-Image2Point: Towards Real-Time Point Cloud Reconstruction of a
Single Image using 3D Supervision
- Authors: AmirHossein Zamani, Amir G. Aghdam and Kamran Ghaffari T
- Abstract summary: A key question in the problem of 3D reconstruction is how to train a machine or a robot to model 3D objects.
This study addresses current problems in reconstructing objects displayed in a single-view image in a faster (real-time) fashion.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A key question in the problem of 3D reconstruction is how to train a machine
or a robot to model 3D objects. Many tasks like navigation in real-time systems
such as autonomous vehicles directly depend on this problem. These systems
usually have limited computational power. Despite considerable progress in 3D
reconstruction systems in recent years, applying them to real-time systems such
as navigation systems in autonomous vehicles is still challenging due to the
high complexity and computational demand of the existing methods. This study
addresses current problems in reconstructing objects displayed in a single-view
image in a faster (real-time) fashion. To this end, a simple yet powerful deep
neural framework is developed. The proposed framework consists of two
components: the feature extractor module and the 3D generator module. We use
point cloud representation for the output of our reconstruction module. The
ShapeNet dataset is utilized to compare the method with the existing results in
terms of computation time and accuracy. Simulations demonstrate the superior
performance of the proposed method.
Index Terms-Real-time 3D reconstruction, single-view reconstruction,
supervised learning, deep neural network
Related papers
- Single-view 3D Mesh Reconstruction for Seen and Unseen Categories [69.29406107513621]
Single-view 3D Mesh Reconstruction is a fundamental computer vision task that aims at recovering 3D shapes from single-view RGB images.
This paper tackles Single-view 3D Mesh Reconstruction, to study the model generalization on unseen categories.
We propose an end-to-end two-stage network, GenMesh, to break the category boundaries in reconstruction.
arXiv Detail & Related papers (2022-08-04T14:13:35Z) - FvOR: Robust Joint Shape and Pose Optimization for Few-view Object
Reconstruction [37.81077373162092]
Reconstructing an accurate 3D object model from a few image observations remains a challenging problem in computer vision.
We present FvOR, a learning-based object reconstruction method that predicts accurate 3D models given a few images with noisy input poses.
arXiv Detail & Related papers (2022-05-16T15:39:27Z) - 3D Reconstruction Using a Linear Laser Scanner and a Camera [5.733401663293044]
This study systematically reviews some basic types of 3D reconstruction technology.
It introduces an easy implementation using a linear laser scanner, a camera, and a turntable.
The accuracy and resolution of the point cloud result are quite satisfying.
arXiv Detail & Related papers (2021-12-01T15:20:24Z) - Deep3DPose: Realtime Reconstruction of Arbitrarily Posed Human Bodies
from Single RGB Images [5.775625085664381]
We introduce an approach that accurately reconstructs 3D human poses and detailed 3D full-body geometric models from single images in realtime.
Key idea of our approach is a novel end-to-end multi-task deep learning framework that uses single images to predict five outputs simultaneously.
We show the system advances the frontier of 3D human body and pose reconstruction from single images by quantitative evaluations and comparisons with state-of-the-art methods.
arXiv Detail & Related papers (2021-06-22T04:26:11Z) - SAIL-VOS 3D: A Synthetic Dataset and Baselines for Object Detection and
3D Mesh Reconstruction from Video Data [124.2624568006391]
We present SAIL-VOS 3D: a synthetic video dataset with frame-by-frame mesh annotations.
We also develop first baselines for reconstruction of 3D meshes from video data via temporal models.
arXiv Detail & Related papers (2021-05-18T15:42:37Z) - An Effective Loss Function for Generating 3D Models from Single 2D Image
without Rendering [0.0]
Differentiable rendering is a very successful technique that applies to a Single-View 3D Reconstruction.
Currents use losses based on pixels between a rendered image of some 3D reconstructed object and ground-truth images from given matched viewpoints to optimise parameters of the 3D shape.
We propose a novel effective loss function that evaluates how well the projections of reconstructed 3D point clouds cover the ground truth object's silhouette.
arXiv Detail & Related papers (2021-03-05T00:02:18Z) - SCFusion: Real-time Incremental Scene Reconstruction with Semantic
Completion [86.77318031029404]
We propose a framework that performs scene reconstruction and semantic scene completion jointly in an incremental and real-time manner.
Our framework relies on a novel neural architecture designed to process occupancy maps and leverages voxel states to accurately and efficiently fuse semantic completion with the 3D global model.
arXiv Detail & Related papers (2020-10-26T15:31:52Z) - Amodal 3D Reconstruction for Robotic Manipulation via Stability and
Connectivity [3.359622001455893]
Learning-based 3D object reconstruction enables single- or few-shot estimation of 3D object models.
Existing 3D reconstruction techniques optimize for visual reconstruction fidelity, typically measured by chamfer distance or voxel IOU.
We propose ARM, an amodal 3D reconstruction system that introduces a stability prior over object shapes, (2) a connectivity prior, and (3) a multi-channel input representation.
arXiv Detail & Related papers (2020-09-28T08:52:54Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z) - Pix2Vox++: Multi-scale Context-aware 3D Object Reconstruction from
Single and Multiple Images [56.652027072552606]
We propose a novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++.
By using a well-designed encoder-decoder, it generates a coarse 3D volume from each input image.
A multi-scale context-aware fusion module is then introduced to adaptively select high-quality reconstructions for different parts from all coarse 3D volumes to obtain a fused 3D volume.
arXiv Detail & Related papers (2020-06-22T13:48:09Z) - From Image Collections to Point Clouds with Self-supervised Shape and
Pose Networks [53.71440550507745]
Reconstructing 3D models from 2D images is one of the fundamental problems in computer vision.
We propose a deep learning technique for 3D object reconstruction from a single image.
We learn both 3D point cloud reconstruction and pose estimation networks in a self-supervised manner.
arXiv Detail & Related papers (2020-05-05T04:25:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.