Neural network with optimal neuron activation functions based on
additive Gaussian process regression
- URL: http://arxiv.org/abs/2301.05567v1
- Date: Fri, 13 Jan 2023 14:19:17 GMT
- Title: Neural network with optimal neuron activation functions based on
additive Gaussian process regression
- Authors: Sergei Manzhos, Manabu Ihara
- Abstract summary: More flexible neuron activation functions would allow using fewer neurons and layers and improve expressive power.
We show that additive Gaussian process regression (GPR) can be used to construct optimal neuron activation functions that are individual to each neuron.
An approach is also introduced that avoids non-linear fitting of neural network parameters.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feed-forward neural networks (NN) are a staple machine learning method widely
used in many areas of science and technology. While even a single-hidden layer
NN is a universal approximator, its expressive power is limited by the use of
simple neuron activation functions (such as sigmoid functions) that are
typically the same for all neurons. More flexible neuron activation functions
would allow using fewer neurons and layers and thereby save computational cost
and improve expressive power. We show that additive Gaussian process regression
(GPR) can be used to construct optimal neuron activation functions that are
individual to each neuron. An approach is also introduced that avoids
non-linear fitting of neural network parameters. The resulting method combines
the advantage of robustness of a linear regression with the higher expressive
power of a NN. We demonstrate the approach by fitting the potential energy
surface of the water molecule. Without requiring any non-linear optimization,
the additive GPR based approach outperforms a conventional NN in the high
accuracy regime, where a conventional NN suffers more from overfitting.
Related papers
- A More Accurate Approximation of Activation Function with Few Spikes Neurons [6.306126887439676]
spiking neural networks (SNNs) have attracted lots of attention as energy-efficient neural networks.
conventional spiking neurons, such as leaky integrate-and-fire neurons, cannot accurately represent complex non-linear activation functions.
arXiv Detail & Related papers (2024-08-19T02:08:56Z) - Fast gradient-free activation maximization for neurons in spiking neural networks [5.805438104063613]
We present a framework with an efficient design for such a loop.
We track changes in the optimal stimuli for artificial neurons during training.
This formation of refined optimal stimuli is associated with an increase in classification accuracy.
arXiv Detail & Related papers (2023-12-28T18:30:13Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Orders-of-coupling representation with a single neural network with
optimal neuron activation functions and without nonlinear parameter
optimization [0.0]
We show that neural network models of orders-of-coupling representations can be easily built by using a recently proposed neural network with optimal neuron activation functions.
Examples are given of representations of molecular potential energy surfaces.
arXiv Detail & Related papers (2023-02-11T06:27:26Z) - Spiking neural network for nonlinear regression [68.8204255655161]
Spiking neural networks carry the potential for a massive reduction in memory and energy consumption.
They introduce temporal and neuronal sparsity, which can be exploited by next-generation neuromorphic hardware.
A framework for regression using spiking neural networks is proposed.
arXiv Detail & Related papers (2022-10-06T13:04:45Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - Improving Spiking Neural Network Accuracy Using Time-based Neurons [0.24366811507669117]
Research on neuromorphic computing systems based on low-power spiking neural networks using analog neurons is in the spotlight.
As technology scales down, analog neurons are difficult to scale, and they suffer from reduced voltage headroom/dynamic range and circuit nonlinearities.
This paper first models the nonlinear behavior of existing current-mirror-based voltage-domain neurons designed in a 28nm process, and show SNN inference accuracy can be severely degraded by the effect of neuron's nonlinearity.
We propose a novel neuron, which processes incoming spikes in the time domain and greatly improves the linearity, thereby improving the inference accuracy compared to the
arXiv Detail & Related papers (2022-01-05T00:24:45Z) - Two-argument activation functions learn soft XOR operations like
cortical neurons [6.88204255655161]
We learn canonical activation functions with two input arguments, analogous to basal and apical dendrites.
Remarkably, the resultant nonlinearities often produce soft XOR functions.
Networks with these nonlinearities learn faster and perform better than conventional ReLU nonlinearities with matched parameter counts.
arXiv Detail & Related papers (2021-10-13T17:06:20Z) - Going Beyond Linear RL: Sample Efficient Neural Function Approximation [76.57464214864756]
We study function approximation with two-layer neural networks.
Our results significantly improve upon what can be attained with linear (or eluder dimension) methods.
arXiv Detail & Related papers (2021-07-14T03:03:56Z) - Flexible Transmitter Network [84.90891046882213]
Current neural networks are mostly built upon the MP model, which usually formulates the neuron as executing an activation function on the real-valued weighted aggregation of signals received from other neurons.
We propose the Flexible Transmitter (FT) model, a novel bio-plausible neuron model with flexible synaptic plasticity.
We present the Flexible Transmitter Network (FTNet), which is built on the most common fully-connected feed-forward architecture.
arXiv Detail & Related papers (2020-04-08T06:55:12Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.