A Dataset-Dispersion Perspective on Reconstruction Versus Recognition in
Single-View 3D Reconstruction Networks
- URL: http://arxiv.org/abs/2111.15158v1
- Date: Tue, 30 Nov 2021 06:33:35 GMT
- Title: A Dataset-Dispersion Perspective on Reconstruction Versus Recognition in
Single-View 3D Reconstruction Networks
- Authors: Yefan Zhou, Yiru Shen, Yujun Yan, Chen Feng, Yaoqing Yang
- Abstract summary: We introduce the dispersion score, a new data-driven metric, to quantify this leading factor and study its effect on NNs.
We show that the proposed metric is a principal way to analyze reconstruction quality and provides novel information in addition to the conventional reconstruction score.
- Score: 16.348294592961327
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Neural networks (NN) for single-view 3D reconstruction (SVR) have gained in
popularity. Recent work points out that for SVR, most cutting-edge NNs have
limited performance on reconstructing unseen objects because they rely
primarily on recognition (i.e., classification-based methods) rather than shape
reconstruction. To understand this issue in depth, we provide a systematic
study on when and why NNs prefer recognition to reconstruction and vice versa.
Our finding shows that a leading factor in determining recognition versus
reconstruction is how dispersed the training data is. Thus, we introduce the
dispersion score, a new data-driven metric, to quantify this leading factor and
study its effect on NNs. We hypothesize that NNs are biased toward recognition
when training images are more dispersed and training shapes are less dispersed.
Our hypothesis is supported and the dispersion score is proved effective
through our experiments on synthetic and benchmark datasets. We show that the
proposed metric is a principal way to analyze reconstruction quality and
provides novel information in addition to the conventional reconstruction
score.
Related papers
- On the Reconstruction of Training Data from Group Invariant Networks [26.3215664733825]
Reconstructing training data from trained neural networks is an active area of research with significant implications for privacy and explainability.
Recent advances have demonstrated the feasibility of this process for several data types.
However, reconstructing data from group-invariant neural networks poses distinct challenges that remain largely unexplored.
arXiv Detail & Related papers (2024-11-25T15:05:00Z) - DCNN: Dual Cross-current Neural Networks Realized Using An Interactive Deep Learning Discriminator for Fine-grained Objects [48.65846477275723]
This study proposes novel dual-current neural networks (DCNN) to improve the accuracy of fine-grained image classification.
The main novel design features for constructing a weakly supervised learning backbone model DCNN include (a) extracting heterogeneous data, (b) keeping the feature map resolution unchanged, (c) expanding the receptive field, and (d) fusing global representations and local features.
arXiv Detail & Related papers (2024-05-07T07:51:28Z) - Understanding and Leveraging the Learning Phases of Neural Networks [7.1169582271841625]
The learning dynamics of deep neural networks are not well understood.
We comprehensively analyze the learning dynamics by investigating a layer's reconstruction ability of the input and prediction performance.
We show the existence of three phases using common datasets and architectures such as ResNet and VGG.
arXiv Detail & Related papers (2023-12-11T23:20:58Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Robust Geometry-Preserving Depth Estimation Using Differentiable
Rendering [93.94371335579321]
We propose a learning framework that trains models to predict geometry-preserving depth without requiring extra data or annotations.
Comprehensive experiments underscore our framework's superior generalization capabilities.
Our innovative loss functions empower the model to autonomously recover domain-specific scale-and-shift coefficients.
arXiv Detail & Related papers (2023-09-18T12:36:39Z) - Neural Poisson Surface Reconstruction: Resolution-Agnostic Shape
Reconstruction from Point Clouds [53.02191521770926]
We introduce Neural Poisson Surface Reconstruction (nPSR), an architecture for shape reconstruction that addresses the challenge of recovering 3D shapes from points.
nPSR exhibits two main advantages: First, it enables efficient training on low-resolution data while achieving comparable performance at high-resolution evaluation.
Overall, the neural Poisson surface reconstruction not only improves upon the limitations of classical deep neural networks in shape reconstruction but also achieves superior results in terms of reconstruction quality, running time, and resolution agnosticism.
arXiv Detail & Related papers (2023-08-03T13:56:07Z) - NeurAR: Neural Uncertainty for Autonomous 3D Reconstruction [64.36535692191343]
Implicit neural representations have shown compelling results in offline 3D reconstruction and also recently demonstrated the potential for online SLAM systems.
This paper addresses two key challenges: 1) seeking a criterion to measure the quality of the candidate viewpoints for the view planning based on the new representations, and 2) learning the criterion from data that can generalize to different scenes instead of hand-crafting one.
Our method demonstrates significant improvements on various metrics for the rendered image quality and the geometry quality of the reconstructed 3D models when compared with variants using TSDF or reconstruction without view planning.
arXiv Detail & Related papers (2022-07-22T10:05:36Z) - The learning phases in NN: From Fitting the Majority to Fitting a Few [2.5991265608180396]
We analyze a layer's reconstruction ability of the input and prediction performance based on the evolution of parameters during training.
We also assess the behavior using common datasets and architectures from computer vision such as ResNet and VGG.
arXiv Detail & Related papers (2022-02-16T19:11:42Z) - Self-Supervised Learning for MRI Reconstruction with a Parallel Network
Training Framework [24.46388892324129]
The proposed method is flexible and can be employed in any existing deep learning-based method.
The effectiveness of the method is evaluated on an open brain MRI dataset.
arXiv Detail & Related papers (2021-09-26T06:09:56Z) - Over-and-Under Complete Convolutional RNN for MRI Reconstruction [57.95363471940937]
Recent deep learning-based methods for MR image reconstruction usually leverage a generic auto-encoder architecture.
We propose an Over-and-Under Complete Convolu?tional Recurrent Neural Network (OUCR), which consists of an overcomplete and an undercomplete Convolutional Recurrent Neural Network(CRNN)
The proposed method achieves significant improvements over the compressed sensing and popular deep learning-based methods with less number of trainable parameters.
arXiv Detail & Related papers (2021-06-16T15:56:34Z) - Procrustean Regression Networks: Learning 3D Structure of Non-Rigid
Objects from 2D Annotations [42.476537776831314]
We propose a novel framework for training neural networks which is capable of learning 3D information of non-rigid objects.
The proposed framework shows superior reconstruction performance to the state-of-the-art method on the Human 3.6M, 300-VW, and SURREAL datasets.
arXiv Detail & Related papers (2020-07-21T17:29:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.