Neuron-Specific Dropout: A Deterministic Regularization Technique to
Prevent Neural Networks from Overfitting & Reduce Dependence on Large
Training Samples
- URL: http://arxiv.org/abs/2201.06938v1
- Date: Thu, 13 Jan 2022 13:10:30 GMT
- Title: Neuron-Specific Dropout: A Deterministic Regularization Technique to
Prevent Neural Networks from Overfitting & Reduce Dependence on Large
Training Samples
- Authors: Joshua Shunk
- Abstract summary: NSDropout looks at both the training pass, and validation pass, of a layer in a model.
By comparing the average values produced by each neuron for each class in a data set, the network is able to drop targeted units.
The layer is able to predict what features, or noise, the model is looking at during testing that isn't present when looking at samples from validation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In order to develop complex relationships between their inputs and outputs,
deep neural networks train and adjust large number of parameters. To make these
networks work at high accuracy, vast amounts of data are needed. Sometimes,
however, the quantity of data needed is not present or obtainable for training.
Neuron-specific dropout (NSDropout) is a tool to address this problem.
NSDropout looks at both the training pass, and validation pass, of a layer in a
model. By comparing the average values produced by each neuron for each class
in a data set, the network is able to drop targeted units. The layer is able to
predict what features, or noise, the model is looking at during testing that
isn't present when looking at samples from validation. Unlike dropout, the
"thinned" networks cannot be "unthinned" for testing. Neuron-specific dropout
has proved to achieve similar, if not better, testing accuracy with far less
data than traditional methods including dropout and other regularization
methods. Experimentation has shown that neuron-specific dropout reduces the
chance of a network overfitting and reduces the need for large training samples
on supervised learning tasks in image recognition, all while producing
best-in-class results.
Related papers
- NeuFair: Neural Network Fairness Repair with Dropout [19.49034966552718]
This paper investigates neuron dropout as a post-processing bias mitigation for deep neural networks (DNNs)
We show that our design of randomized algorithms is effective and efficient in improving fairness (up to 69%) with minimal or no model performance degradation.
arXiv Detail & Related papers (2024-07-05T05:45:34Z) - Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - When do Convolutional Neural Networks Stop Learning? [0.0]
Convolutional Neural Networks (CNNs) have demonstrated outstanding performance in computer vision tasks.
Current practice is to stop training when the training loss decreases and the gap between training and validation error increases.
This research work introduces a hypothesis that analyzes the data variation across all the layers of a CNN variant to anticipate its near-optimal learning capacity.
arXiv Detail & Related papers (2024-03-04T20:35:09Z) - Sampling weights of deep neural networks [1.2370077627846041]
We introduce a probability distribution, combined with an efficient sampling algorithm, for weights and biases of fully-connected neural networks.
In a supervised learning context, no iterative optimization or gradient computations of internal network parameters are needed.
We prove that sampled networks are universal approximators.
arXiv Detail & Related papers (2023-06-29T10:13:36Z) - Benign Overfitting for Two-layer ReLU Convolutional Neural Networks [60.19739010031304]
We establish algorithm-dependent risk bounds for learning two-layer ReLU convolutional neural networks with label-flipping noise.
We show that, under mild conditions, the neural network trained by gradient descent can achieve near-zero training loss and Bayes optimal test risk.
arXiv Detail & Related papers (2023-03-07T18:59:38Z) - Dropout Reduces Underfitting [85.61466286688385]
In this study, we demonstrate that dropout can also mitigate underfitting when used at the start of training.
We find dropout reduces the directional variance of gradients across mini-batches and helps align the mini-batch gradients with the entire dataset's gradient.
Our findings lead us to a solution for improving performance in underfitting models - early dropout: dropout is applied only during the initial phases of training, and turned off afterwards.
arXiv Detail & Related papers (2023-03-02T18:59:15Z) - DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning [5.2319020651074215]
We propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP)
Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution.
We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors.
arXiv Detail & Related papers (2023-02-25T08:16:21Z) - Boosted Dynamic Neural Networks [53.559833501288146]
A typical EDNN has multiple prediction heads at different layers of the network backbone.
To optimize the model, these prediction heads together with the network backbone are trained on every batch of training data.
Treating training and testing inputs differently at the two phases will cause the mismatch between training and testing data distributions.
We formulate an EDNN as an additive model inspired by gradient boosting, and propose multiple training techniques to optimize the model effectively.
arXiv Detail & Related papers (2022-11-30T04:23:12Z) - Contextual Dropout: An Efficient Sample-Dependent Dropout Module [60.63525456640462]
Dropout has been demonstrated as a simple and effective module to regularize the training process of deep neural networks.
We propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module.
Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.
arXiv Detail & Related papers (2021-03-06T19:30:32Z) - Advanced Dropout: A Model-free Methodology for Bayesian Dropout
Optimization [62.8384110757689]
Overfitting ubiquitously exists in real-world applications of deep neural networks (DNNs)
The advanced dropout technique applies a model-free and easily implemented distribution with parametric prior, and adaptively adjusts dropout rate.
We evaluate the effectiveness of the advanced dropout against nine dropout techniques on seven computer vision datasets.
arXiv Detail & Related papers (2020-10-11T13:19:58Z) - MaxDropout: Deep Neural Network Regularization Based on Maximum Output
Values [0.0]
MaxDropout is a regularizer for deep neural network models that works in a supervised fashion by removing prominent neurons.
We show that it is possible to improve existing neural networks and provide better results in neural networks when Dropout is replaced by MaxDropout.
arXiv Detail & Related papers (2020-07-27T17:55:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.