Squashing activation functions in benchmark tests: towards eXplainable
Artificial Intelligence using continuous-valued logic
- URL: http://arxiv.org/abs/2010.08760v1
- Date: Sat, 17 Oct 2020 10:42:40 GMT
- Title: Squashing activation functions in benchmark tests: towards eXplainable
Artificial Intelligence using continuous-valued logic
- Authors: Daniel Zeltner, Benedikt Schmid, Gabor Csiszar, Orsolya Csiszar
- Abstract summary: This work demonstrates the first benchmark tests that measure the performance of Squashing functions in neural networks.
Three experiments were carried out to examine their usability and a comparison with the most popular activation functions was made for five different network types.
Results indicate that due to the embedded nilpotent logical operators and the differentiability of the Squashing function, it is possible to solve classification problems.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the past few years, deep neural networks have shown excellent results in
multiple tasks, however, there is still an increasing need to address the
problem of interpretability to improve model transparency, performance, and
safety. Achieving eXplainable Artificial Intelligence (XAI) by combining neural
networks with continuous logic and multi-criteria decision-making tools is one
of the most promising ways to approach this problem: by this combination, the
black-box nature of neural models can be reduced. The continuous logic-based
neural model uses so-called Squashing activation functions, a parametric family
of functions that satisfy natural invariance requirements and contain rectified
linear units as a particular case. This work demonstrates the first benchmark
tests that measure the performance of Squashing functions in neural networks.
Three experiments were carried out to examine their usability and a comparison
with the most popular activation functions was made for five different network
types. The performance was determined by measuring the accuracy, loss, and time
per epoch. These experiments and the conducted benchmarks have proven that the
use of Squashing functions is possible and similar in performance to
conventional activation functions. Moreover, a further experiment was conducted
by implementing nilpotent logical gates to demonstrate how simple
classification tasks can be solved successfully and with high performance. The
results indicate that due to the embedded nilpotent logical operators and the
differentiability of the Squashing function, it is possible to solve
classification problems, where other commonly used activation functions fail.
Related papers
- Adaptive Activation Functions for Predictive Modeling with Sparse
Experimental Data [2.012425476229879]
This study investigates the influence of adaptive or trainable activation functions on classification accuracy and predictive uncertainty in settings characterized by limited data availability.
Our investigation reveals that adaptive activation functions, such as Exponential Linear Unit (ELU) and Softplus, with individual trainable parameters, result in accurate and confident prediction models.
arXiv Detail & Related papers (2024-02-08T04:35:09Z) - The limitation of neural nets for approximation and optimization [0.0]
We are interested in assessing the use of neural networks as surrogate models to approximate and minimize objective functions in optimization problems.
Our study begins by determining the best activation function for approximating the objective functions of popular nonlinear optimization test problems.
arXiv Detail & Related papers (2023-11-21T00:21:15Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Data-aware customization of activation functions reduces neural network
error [0.35172332086962865]
We show that data-aware customization of activation functions can result in striking reductions in neural network error.
A simple substitution with the seagull'' activation function in an already-refined neural network can lead to an order-of-magnitude reduction in error.
arXiv Detail & Related papers (2023-01-16T23:38:37Z) - Stochastic Adaptive Activation Function [1.9199289015460212]
This study proposes a simple yet effective activation function that facilitates different thresholds and adaptive activations according to the positions of units and the contexts of inputs.
Experimental analysis demonstrates that our activation function can provide the benefits of more accurate prediction and earlier convergence in many deep learning applications.
arXiv Detail & Related papers (2022-10-21T01:57:25Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Recognizing and Verifying Mathematical Equations using Multiplicative
Differential Neural Units [86.9207811656179]
We show that memory-augmented neural networks (NNs) can achieve higher-order, memory-augmented extrapolation, stable performance, and faster convergence.
Our models achieve a 1.53% average improvement over current state-of-the-art methods in equation verification and achieve a 2.22% Top-1 average accuracy and 2.96% Top-5 average accuracy for equation completion.
arXiv Detail & Related papers (2021-04-07T03:50:11Z) - A Use of Even Activation Functions in Neural Networks [0.35172332086962865]
We propose an alternative approach to integrate existing knowledge or hypotheses of data structure by constructing custom activation functions.
We show that using an even activation function in one of the fully connected layers improves neural network performance.
arXiv Detail & Related papers (2020-11-23T20:33:13Z) - UNIPoint: Universally Approximating Point Processes Intensities [125.08205865536577]
We provide a proof that a class of learnable functions can universally approximate any valid intensity function.
We implement UNIPoint, a novel neural point process model, using recurrent neural networks to parameterise sums of basis function upon each event.
arXiv Detail & Related papers (2020-07-28T09:31:56Z) - Towards Efficient Processing and Learning with Spikes: New Approaches
for Multi-Spike Learning [59.249322621035056]
We propose two new multi-spike learning rules which demonstrate better performance over other baselines on various tasks.
In the feature detection task, we re-examine the ability of unsupervised STDP with its limitations being presented.
Our proposed learning rules can reliably solve the task over a wide range of conditions without specific constraints being applied.
arXiv Detail & Related papers (2020-05-02T06:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.