Testing for Normality with Neural Networks
- URL: http://arxiv.org/abs/2009.13831v2
- Date: Wed, 7 Oct 2020 07:47:22 GMT
- Title: Testing for Normality with Neural Networks
- Authors: Milo\v{s} Simi\'c
- Abstract summary: We construct a feedforward neural network that can successfully detect normal distributions by inspecting small samples from them.
The network's accuracy was higher than 96% on a set of larger samples with 250-1000 elements.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we treat the problem of testing for normality as a binary
classification problem and construct a feedforward neural network that can
successfully detect normal distributions by inspecting small samples from them.
The numerical experiments conducted on small samples with no more than 100
elements indicated that the neural network which we trained was more accurate
and far more powerful than the most frequently used and most powerful standard
tests of normality: Shapiro-Wilk, Anderson-Darling, Lilliefors and
Jarque-Berra, as well as the kernel tests of goodness-of-fit. The neural
network had the AUROC score of almost 1, which corresponds to the perfect
binary classifier. Additionally, the network's accuracy was higher than 96% on
a set of larger samples with 250-1000 elements. Since the normality of data is
an assumption of numerous techniques for analysis and inference, the neural
network constructed in this study has a very high potential for use in everyday
practice of statistics, data analysis and machine learning in both science and
industry.
Related papers
- Residual Random Neural Networks [0.0]
Single-layer feedforward neural network with random weights is a recurring motif in the neural networks literature.
We show that one can obtain good classification results even if the number of hidden neurons has the same order of magnitude as the dimensionality of the data samples.
arXiv Detail & Related papers (2024-10-25T22:00:11Z) - Training Guarantees of Neural Network Classification Two-Sample Tests by Kernel Analysis [58.435336033383145]
We construct and analyze a neural network two-sample test to determine whether two datasets came from the same distribution.
We derive the theoretical minimum training time needed to ensure the NTK two-sample test detects a deviation-level between the datasets.
We show that the statistical power associated with the neural network two-sample test goes to 1 as the neural network training samples and test evaluation samples go to infinity.
arXiv Detail & Related papers (2024-07-05T18:41:16Z) - How many samples are needed to train a deep neural network? [6.042269506496206]
We study the question of how much data are needed to train a ReLU feed-forward neural network.
Our results underpin the common belief that neural networks need "many" training samples.
arXiv Detail & Related papers (2024-05-26T21:07:21Z) - Verified Neural Compressed Sensing [58.98637799432153]
We develop the first (to the best of our knowledge) provably correct neural networks for a precise computational task.
We show that for modest problem dimensions (up to 50), we can train neural networks that provably recover a sparse vector from linear and binarized linear measurements.
We show that the complexity of the network can be adapted to the problem difficulty and solve problems where traditional compressed sensing methods are not known to provably work.
arXiv Detail & Related papers (2024-05-07T12:20:12Z) - Learning from higher-order statistics, efficiently: hypothesis tests, random features, and neural networks [7.503293567983987]
We study how efficient are neural networks at extracting features from higher-order cumulants.
We show that neural networks do indeed learn to distinguish the two distributions with quadratic sample complexity.
Our results show that neural networks extract information from higher-ordercorrelations in the spiked cumulant model efficiently.
arXiv Detail & Related papers (2023-12-22T18:55:25Z) - Sampling weights of deep neural networks [1.2370077627846041]
We introduce a probability distribution, combined with an efficient sampling algorithm, for weights and biases of fully-connected neural networks.
In a supervised learning context, no iterative optimization or gradient computations of internal network parameters are needed.
We prove that sampled networks are universal approximators.
arXiv Detail & Related papers (2023-06-29T10:13:36Z) - Diffused Redundancy in Pre-trained Representations [98.55546694886819]
We take a closer look at how features are encoded in pre-trained representations.
We find that learned representations in a given layer exhibit a degree of diffuse redundancy.
Our findings shed light on the nature of representations learned by pre-trained deep neural networks.
arXiv Detail & Related papers (2023-05-31T21:00:50Z) - Benign Overfitting for Two-layer ReLU Convolutional Neural Networks [60.19739010031304]
We establish algorithm-dependent risk bounds for learning two-layer ReLU convolutional neural networks with label-flipping noise.
We show that, under mild conditions, the neural network trained by gradient descent can achieve near-zero training loss and Bayes optimal test risk.
arXiv Detail & Related papers (2023-03-07T18:59:38Z) - Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity
on Pruned Neural Networks [79.74580058178594]
We analyze the performance of training a pruned neural network by analyzing the geometric structure of the objective function.
We show that the convex region near a desirable model with guaranteed generalization enlarges as the neural network model is pruned.
arXiv Detail & Related papers (2021-10-12T01:11:07Z) - Dive into Layers: Neural Network Capacity Bounding using Algebraic
Geometry [55.57953219617467]
We show that the learnability of a neural network is directly related to its size.
We use Betti numbers to measure the topological geometric complexity of input data and the neural network.
We perform the experiments on a real-world dataset MNIST and the results verify our analysis and conclusion.
arXiv Detail & Related papers (2021-09-03T11:45:51Z) - Statistical Guarantees for Regularized Neural Networks [4.254099382808598]
We develop a general statistical guarantee for estimators that consist of a least-squares term and a regularizer.
Our results establish a mathematical basis for regularized estimation of neural networks.
arXiv Detail & Related papers (2020-05-30T15:28:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.