3D-NVS: A 3D Supervision Approach for Next View Selection
- URL: http://arxiv.org/abs/2012.01743v1
- Date: Thu, 3 Dec 2020 07:50:16 GMT
- Title: 3D-NVS: A 3D Supervision Approach for Next View Selection
- Authors: Kumar Ashutosh, Saurabh Kumar, Subhasis Chaudhuri
- Abstract summary: We present a classification based approach for the next best view selection.
We show how we can plausibly obtain a supervisory signal for this task.
- Score: 22.662440687566587
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a classification based approach for the next best view selection
and show how we can plausibly obtain a supervisory signal for this task. The
proposed approach is end-to-end trainable and aims to get the best possible 3D
reconstruction quality with a pair of passively acquired 2D views. The proposed
model consists of two stages: a classifier and a reconstructor network trained
jointly via the indirect 3D supervision from ground truth voxels. While
testing, the proposed method assumes no prior knowledge of the underlying 3D
shape for selecting the next best view. We demonstrate the proposed method's
effectiveness via detailed experiments on synthetic and real images and show
how it provides improved reconstruction quality than the existing state of the
art 3D reconstruction and the next best view prediction techniques.
Related papers
- Gaussian Splatting: 3D Reconstruction and Novel View Synthesis, a Review [0.08823202672546056]
This review paper focuses on state-of-the-art techniques for 3D reconstruction, including the generation of novel, unseen views.
An overview of recent developments in the Gaussian Splatting method is provided, covering input types, model structures, output representations, and training strategies.
arXiv Detail & Related papers (2024-05-06T12:32:38Z) - RadOcc: Learning Cross-Modality Occupancy Knowledge through Rendering
Assisted Distillation [50.35403070279804]
3D occupancy prediction is an emerging task that aims to estimate the occupancy states and semantics of 3D scenes using multi-view images.
We propose RadOcc, a Rendering assisted distillation paradigm for 3D Occupancy prediction.
arXiv Detail & Related papers (2023-12-19T03:39:56Z) - Take-A-Photo: 3D-to-2D Generative Pre-training of Point Cloud Models [97.58685709663287]
generative pre-training can boost the performance of fundamental models in 2D vision.
In 3D vision, the over-reliance on Transformer-based backbones and the unordered nature of point clouds have restricted the further development of generative pre-training.
We propose a novel 3D-to-2D generative pre-training method that is adaptable to any point cloud model.
arXiv Detail & Related papers (2023-07-27T16:07:03Z) - 3D-LatentMapper: View Agnostic Single-View Reconstruction of 3D Shapes [0.0]
We propose a novel framework that leverages the intermediate latent spaces of Vision Transformer (ViT) and a joint image-text representational model, CLIP, for fast and efficient Single View Reconstruction (SVR)
We use the ShapeNetV2 dataset and perform extensive experiments with comparisons to SOTA methods to demonstrate our method's effectiveness.
arXiv Detail & Related papers (2022-12-05T11:45:26Z) - Uncertainty-Driven Active Vision for Implicit Scene Reconstruction [31.890470553111122]
We propose an uncertainty-driven active vision approach for implicit scene reconstruction.
We develop an occupancy-based reconstruction method which accurately represents scenes using either 2D or 3D supervision.
arXiv Detail & Related papers (2022-10-03T14:45:54Z) - Uncertainty Guided Policy for Active Robotic 3D Reconstruction using
Neural Radiance Fields [82.21033337949757]
This paper introduces a ray-based volumetric uncertainty estimator, which computes the entropy of the weight distribution of the color samples along each ray of the object's implicit neural representation.
We show that it is possible to infer the uncertainty of the underlying 3D geometry given a novel view with the proposed estimator.
We present a next-best-view selection policy guided by the ray-based volumetric uncertainty in neural radiance fields-based representations.
arXiv Detail & Related papers (2022-09-17T21:28:57Z) - A Simple Baseline for Multi-Camera 3D Object Detection [94.63944826540491]
3D object detection with surrounding cameras has been a promising direction for autonomous driving.
We present SimMOD, a Simple baseline for Multi-camera Object Detection.
We conduct extensive experiments on the 3D object detection benchmark of nuScenes to demonstrate the effectiveness of SimMOD.
arXiv Detail & Related papers (2022-08-22T03:38:01Z) - NeurAR: Neural Uncertainty for Autonomous 3D Reconstruction [64.36535692191343]
Implicit neural representations have shown compelling results in offline 3D reconstruction and also recently demonstrated the potential for online SLAM systems.
This paper addresses two key challenges: 1) seeking a criterion to measure the quality of the candidate viewpoints for the view planning based on the new representations, and 2) learning the criterion from data that can generalize to different scenes instead of hand-crafting one.
Our method demonstrates significant improvements on various metrics for the rendered image quality and the geometry quality of the reconstructed 3D models when compared with variants using TSDF or reconstruction without view planning.
arXiv Detail & Related papers (2022-07-22T10:05:36Z) - Improving 3D Object Detection with Channel-wise Transformer [58.668922561622466]
We propose a two-stage 3D object detection framework (CT3D) with minimal hand-crafted design.
CT3D simultaneously performs proposal-aware embedding and channel-wise context aggregation.
It achieves the AP of 81.77% in the moderate car category on the KITTI test 3D detection benchmark.
arXiv Detail & Related papers (2021-08-23T02:03:40Z) - Pose Estimation and 3D Reconstruction of Vehicles from Stereo-Images
Using a Subcategory-Aware Shape Prior [0.0]
3D reconstruction of objects or a computer vision is a prerequisite for many applications such as mobile robotics autonomous driving.
The goal of this paper is to show how 3D object reconstruction can profit from prior shape observations.
arXiv Detail & Related papers (2021-07-22T19:47:49Z) - Next-best-view Regression using a 3D Convolutional Neural Network [0.9449650062296823]
We propose a data-driven approach to address the next-best-view problem.
The proposed approach trains a 3D convolutional neural network with previous reconstructions in order to regress the btxtposition of the next-best-view.
We have validated the proposed approach making use of two groups of experiments.
arXiv Detail & Related papers (2021-01-23T01:50:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.