Neural Random Projection: From the Initial Task To the Input Similarity
Problem
- URL: http://arxiv.org/abs/2010.04555v1
- Date: Fri, 9 Oct 2020 13:20:24 GMT
- Title: Neural Random Projection: From the Initial Task To the Input Similarity
Problem
- Authors: Alan Savushkin, Nikita Benkovich and Dmitry Golubev
- Abstract summary: We propose a novel approach for implicit data representation to evaluate similarity of input data using a trained neural network.
The proposed technique explicitly takes into account the initial task and significantly reduces the size of the vector representation.
Our experimental results show that the proposed approach achieves competitive results on the input similarity task.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a novel approach for implicit data representation
to evaluate similarity of input data using a trained neural network. In
contrast to the previous approach, which uses gradients for representation, we
utilize only the outputs from the last hidden layer of a neural network and do
not use a backward step. The proposed technique explicitly takes into account
the initial task and significantly reduces the size of the vector
representation, as well as the computation time. The key point is minimization
of information loss between layers. Generally, a neural network discards
information that is not related to the problem, which makes the last hidden
layer representation useless for input similarity task. In this work, we
consider two main causes of information loss: correlation between neurons and
insufficient size of the last hidden layer. To reduce the correlation between
neurons we use orthogonal weight initialization for each layer and modify the
loss function to ensure orthogonality of the weights during training. Moreover,
we show that activation functions can potentially increase correlation. To
solve this problem, we apply modified Batch-Normalization with Dropout. Using
orthogonal weight matrices allow us to consider such neural networks as an
application of the Random Projection method and get a lower bound estimate for
the size of the last hidden layer. We perform experiments on MNIST and physical
examination datasets. In both experiments, initially, we split a set of labels
into two disjoint subsets to train a neural network for binary classification
problem, and then use this model to measure similarity between input data and
define hidden classes. Our experimental results show that the proposed approach
achieves competitive results on the input similarity task while reducing both
computation time and the size of the input representation.
Related papers
- RelChaNet: Neural Network Feature Selection using Relative Change Scores [0.0]
We introduce RelChaNet, a novel and lightweight feature selection algorithm that uses neuron pruning and regrowth in the input layer of a dense neural network.
Our approach generally outperforms the current state-of-the-art methods, and in particular improves the average accuracy by 2% on the MNIST dataset.
arXiv Detail & Related papers (2024-10-03T09:56:39Z) - An Initialization Schema for Neuronal Networks on Tabular Data [0.9155684383461983]
We show that a binomial neural network can be used effectively on tabular data.
The proposed approach shows a simple but effective approach for initializing the first hidden layer in neural networks.
We evaluate our approach on multiple public datasets and showcase the improved performance compared to other neural network-based approaches.
arXiv Detail & Related papers (2023-11-07T13:52:35Z) - Using Linear Regression for Iteratively Training Neural Networks [4.873362301533824]
We present a simple linear regression based approach for learning the weights and biases of a neural network.
The approach is intended to be to larger, more complex architectures.
arXiv Detail & Related papers (2023-07-11T11:53:25Z) - Hidden Classification Layers: Enhancing linear separability between
classes in neural networks layers [0.0]
We investigate the impact on deep network performances of a training approach.
We propose a neural network architecture which induces an error function involving the outputs of all the network layers.
arXiv Detail & Related papers (2023-06-09T10:52:49Z) - Diffused Redundancy in Pre-trained Representations [98.55546694886819]
We take a closer look at how features are encoded in pre-trained representations.
We find that learned representations in a given layer exhibit a degree of diffuse redundancy.
Our findings shed light on the nature of representations learned by pre-trained deep neural networks.
arXiv Detail & Related papers (2023-05-31T21:00:50Z) - Cross-Model Comparative Loss for Enhancing Neuronal Utility in Language
Understanding [82.46024259137823]
We propose a cross-model comparative loss for a broad range of tasks.
We demonstrate the universal effectiveness of comparative loss through extensive experiments on 14 datasets from 3 distinct NLU tasks.
arXiv Detail & Related papers (2023-01-10T03:04:27Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - Adaptive conversion of real-valued input into spike trains [91.3755431537592]
This paper presents a biologically plausible method for converting real-valued input into spike trains for processing with spiking neural networks.
The proposed method mimics the adaptive behaviour of retinal ganglion cells and allows input neurons to adapt their response to changes in the statistics of the input.
arXiv Detail & Related papers (2021-04-12T12:33:52Z) - Deep Convolutional Neural Network-based Bernoulli Heatmap for Head Pose
Estimation [11.676928225717337]
This paper proposes a novel Bernoulli heatmap for head pose estimation from a single RGB image.
Our method can achieve the positioning of the head area while estimating the angles of the head.
A deep convolutional neural network (CNN) structure with multiscale representations is adopted to maintain high-resolution information.
arXiv Detail & Related papers (2020-05-24T15:36:29Z) - Revisiting Initialization of Neural Networks [72.24615341588846]
We propose a rigorous estimation of the global curvature of weights across layers by approximating and controlling the norm of their Hessian matrix.
Our experiments on Word2Vec and the MNIST/CIFAR image classification tasks confirm that tracking the Hessian norm is a useful diagnostic tool.
arXiv Detail & Related papers (2020-04-20T18:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.