Multiple View Generation and Classification of Mid-wave Infrared Images
using Deep Learning
- URL: http://arxiv.org/abs/2008.07714v1
- Date: Tue, 18 Aug 2020 02:58:21 GMT
- Title: Multiple View Generation and Classification of Mid-wave Infrared Images
using Deep Learning
- Authors: Maliha Arif, Abhijit Mahalanobis
- Abstract summary: We propose a novel study of generating unseen arbitrary viewpoints for infrared imagery in the non-linear feature subspace.
Our approach on the contrary understands the semantic information in natural images and encapsulates it such that our predicted unseen views possess good 3D representations.
- Score: 1.52292571922932
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel study of generating unseen arbitrary viewpoints for
infrared imagery in the non-linear feature subspace . Current methods use
synthetic images and often result in blurry and distorted outputs. Our approach
on the contrary understands the semantic information in natural images and
encapsulates it such that our predicted unseen views possess good 3D
representations. We further explore the non-linear feature subspace and
conclude that our network does not operate in the Euclidean subspace but rather
in the Riemannian subspace. It does not learn the geometric transformation for
predicting the position of the pixel in the new image but rather learns the
manifold. To this end, we use t-SNE visualisations to conduct a detailed
analysis of our network and perform classification of generated images as a
low-shot learning task.
Related papers
- HVDistill: Transferring Knowledge from Images to Point Clouds via Unsupervised Hybrid-View Distillation [106.09886920774002]
We present a hybrid-view-based knowledge distillation framework, termed HVDistill, to guide the feature learning of a point cloud neural network.
Our method achieves consistent improvements over the baseline trained from scratch and significantly out- performs the existing schemes.
arXiv Detail & Related papers (2024-03-18T14:18:08Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - A domain adaptive deep learning solution for scanpath prediction of
paintings [66.46953851227454]
This paper focuses on the eye-movement analysis of viewers during the visual experience of a certain number of paintings.
We introduce a new approach to predicting human visual attention, which impacts several cognitive functions for humans.
The proposed new architecture ingests images and returns scanpaths, a sequence of points featuring a high likelihood of catching viewers' attention.
arXiv Detail & Related papers (2022-09-22T22:27:08Z) - Decision boundaries and convex hulls in the feature space that deep
learning functions learn from images [0.0]
We study the properties of a low-dimensional manifold that models extract and learn from images.
For image classification models, their last hidden layer is the one where images of each class is separated from other classes and it also has the least number of features.
We observe that geometric arrangements of decision boundaries in feature space is significantly different compared to pixel space.
arXiv Detail & Related papers (2022-02-05T15:09:51Z) - LOLNeRF: Learn from One Look [22.771493686755544]
We present a method for learning a generative 3D model based on neural radiance fields.
We show that, unlike existing methods, one does not need multi-view data to achieve this goal.
arXiv Detail & Related papers (2021-11-19T01:20:01Z) - The Intrinsic Dimension of Images and Its Impact on Learning [60.811039723427676]
It is widely believed that natural image data exhibits low-dimensional structure despite the high dimensionality of conventional pixel representations.
In this work, we apply dimension estimation tools to popular datasets and investigate the role of low-dimensional structure in deep learning.
arXiv Detail & Related papers (2021-04-18T16:29:23Z) - IntroVAC: Introspective Variational Classifiers for Learning
Interpretable Latent Subspaces [6.574517227976925]
IntroVAC learns interpretable latent subspaces by exploiting information from an additional label.
We show that IntroVAC is able to learn meaningful directions in the latent space enabling fine manipulation of image attributes.
arXiv Detail & Related papers (2020-08-03T10:21:41Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z) - Learning Depth With Very Sparse Supervision [57.911425589947314]
This paper explores the idea that perception gets coupled to 3D properties of the world via interaction with the environment.
We train a specialized global-local network architecture with what would be available to a robot interacting with the environment.
Experiments on several datasets show that, when ground truth is available even for just one of the image pixels, the proposed network can learn monocular dense depth estimation up to 22.5% more accurately than state-of-the-art approaches.
arXiv Detail & Related papers (2020-03-02T10:44:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.