Neuroevolutionary algorithms driven by neuron coverage metrics for
semi-supervised classification
- URL: http://arxiv.org/abs/2303.02801v1
- Date: Sun, 5 Mar 2023 23:38:44 GMT
- Title: Neuroevolutionary algorithms driven by neuron coverage metrics for
semi-supervised classification
- Authors: Roberto Santana, Ivan Hidalgo-Cenalmor, Unai Garciarena, Alexander
Mendiburu, Jose Antonio Lozano
- Abstract summary: In some machine learning applications the availability of labeled instances for supervised classification is limited while unlabeled instances are abundant.
We introduce neuroevolutionary approaches that exploit unlabeled instances by using neuron coverage metrics computed on the neural network architecture encoded by each candidate solution.
- Score: 60.60571130467197
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In some machine learning applications the availability of labeled instances
for supervised classification is limited while unlabeled instances are
abundant. Semi-supervised learning algorithms deal with these scenarios and
attempt to exploit the information contained in the unlabeled examples. In this
paper, we address the question of how to evolve neural networks for
semi-supervised problems. We introduce neuroevolutionary approaches that
exploit unlabeled instances by using neuron coverage metrics computed on the
neural network architecture encoded by each candidate solution. Neuron coverage
metrics resemble code coverage metrics used to test software, but are oriented
to quantify how the different neural network components are covered by test
instances. In our neuroevolutionary approach, we define fitness functions that
combine classification accuracy computed on labeled examples and neuron
coverage metrics evaluated using unlabeled examples. We assess the impact of
these functions on semi-supervised problems with a varying amount of labeled
instances. Our results show that the use of neuron coverage metrics helps
neuroevolution to become less sensitive to the scarcity of labeled data, and
can lead in some cases to a more robust generalization of the learned
classifiers.
Related papers
- Benign Overfitting for Two-layer ReLU Convolutional Neural Networks [60.19739010031304]
We establish algorithm-dependent risk bounds for learning two-layer ReLU convolutional neural networks with label-flipping noise.
We show that, under mild conditions, the neural network trained by gradient descent can achieve near-zero training loss and Bayes optimal test risk.
arXiv Detail & Related papers (2023-03-07T18:59:38Z) - Measures of Information Reflect Memorization Patterns [53.71420125627608]
We show that the diversity in the activation patterns of different neurons is reflective of model generalization and memorization.
Importantly, we discover that information organization points to the two forms of memorization, even for neural activations computed on unlabelled in-distribution examples.
arXiv Detail & Related papers (2022-10-17T20:15:24Z) - Revisiting Gaussian Neurons for Online Clustering with Unknown Number of
Clusters [0.0]
A novel local learning rule is presented that performs online clustering with a maximum limit of the number of cluster to be found.
The experimental results demonstrate stability in the learned parameters across a large number of training samples.
arXiv Detail & Related papers (2022-05-02T14:01:40Z) - Fault-Tolerant Neural Networks from Biological Error Correction Codes [45.82537918529782]
In the grid cells of the mammalian cortex, analog error correction codes have been observed to protect states against neural spiking noise.
Here, we use these biological error correction codes to develop a universal fault-tolerant neural network that achieves reliable computation if the faultiness of each neuron lies below a sharp threshold.
The discovery of a phase transition from faulty to fault-tolerant neural computation suggests a mechanism for reliable computation in the cortex.
arXiv Detail & Related papers (2022-02-25T18:55:46Z) - Regularization by Misclassification in ReLU Neural Networks [3.288086999241324]
We study the implicit bias of ReLU neural networks trained by a variant of SGD where at each step, the label is changed with probability $p$ to a random label.
We show that label noise propels the network to a sparse solution in the following sense: for a typical input, a small fraction of neurons are active, and the firing pattern of the hidden layers is sparser.
arXiv Detail & Related papers (2021-11-03T11:42:38Z) - And/or trade-off in artificial neurons: impact on adversarial robustness [91.3755431537592]
Presence of sufficient number of OR-like neurons in a network can lead to classification brittleness and increased vulnerability to adversarial attacks.
We define AND-like neurons and propose measures to increase their proportion in the network.
Experimental results on the MNIST dataset suggest that our approach holds promise as a direction for further exploration.
arXiv Detail & Related papers (2021-02-15T08:19:05Z) - SeReNe: Sensitivity based Regularization of Neurons for Structured
Sparsity in Neural Networks [13.60023740064471]
SeReNe is a method for learning sparse topologies with a structure.
We define the sensitivity of a neuron as the variation of the network output.
By including the neuron sensitivity in the cost function as a regularization term, we areable to prune neurons with low sensitivity.
arXiv Detail & Related papers (2021-02-07T10:53:30Z) - The Neural Coding Framework for Learning Generative Models [91.0357317238509]
We propose a novel neural generative model inspired by the theory of predictive processing in the brain.
In a similar way, artificial neurons in our generative model predict what neighboring neurons will do, and adjust their parameters based on how well the predictions matched reality.
arXiv Detail & Related papers (2020-12-07T01:20:38Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.