Investigating Continuous Learning in Spiking Neural Networks
- URL: http://arxiv.org/abs/2310.05343v1
- Date: Mon, 9 Oct 2023 02:08:18 GMT
- Title: Investigating Continuous Learning in Spiking Neural Networks
- Authors: C. Tanner Fredieu
- Abstract summary: Third-generation machine learning, also known as spiking neural network architecture, for continuous learning was investigated and compared to conventional models.
All models were able to correctly identify the current classes, but they would immediately see a sharp performance drop in previous classes due to catastrophic forgetting.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, the use of third-generation machine learning, also known as
spiking neural network architecture, for continuous learning was investigated
and compared to conventional models. The experimentation was divided into three
separate phases. The first phase focused on training the conventional models
via transfer learning. The second phase trains a Nengo model from their
library. Lastly, each conventional model is converted into a spiking neural
network and trained. Initial results from phase 1 are inline with known
knowledge about continuous learning within current machine learning literature.
All models were able to correctly identify the current classes, but they would
immediately see a sharp performance drop in previous classes due to
catastrophic forgetting. However, the SNN models were able to retain some
information about previous classes. Although many of the previous classes were
still identified as the current trained classes, the output probabilities
showed a higher than normal value to the actual class. This indicates that the
SNN models do have potential to overcome catastrophic forgetting but much work
is still needed.
Related papers
- Singular leaning coefficients and efficiency in learning theory [0.0]
Singular learning models with non-positive Fisher information matrices include neural networks, reduced-rank regression, Boltzmann machines, normal mixture models, and others.
We examine learning coefficients, which indicate the general learning efficiency of deep linear learning models and three-layer neural network models with ReLU units.
arXiv Detail & Related papers (2025-01-22T09:31:02Z) - Text Classification: Neural Networks VS Machine Learning Models VS Pre-trained Models [0.0]
We present a comparison between different techniques to perform text classification.
We take into consideration seven pre-trained models, three standard neural networks and three machine learning models.
For standard neural networks and machine learning models we also compare two embedding techniques: TF-IDF and GloVe.
arXiv Detail & Related papers (2024-12-30T15:44:05Z) - Learning to Continually Learn with the Bayesian Principle [36.75558255534538]
In this work, we adopt the meta-learning paradigm to combine the strong representational power of neural networks and simple statistical models' robustness to forgetting.
Since the neural networks remain fixed during continual learning, they are protected from catastrophic forgetting.
arXiv Detail & Related papers (2024-05-29T04:53:31Z) - Steganographic Capacity of Deep Learning Models [12.974139332068491]
We consider the steganographic capacity of several learning models.
We train a Multilayer Perceptron (MLP), Convolutional Neural Network (CNN), and Transformer model on a challenging malware classification problem.
We find that the steganographic capacity of the learning models tested is surprisingly high, and that in each case, there is a clear threshold after which model performance rapidly degrades.
arXiv Detail & Related papers (2023-06-25T13:43:35Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - CQural: A Novel CNN based Hybrid Architecture for Quantum Continual
Machine Learning [0.0]
We show that it is possible to circumvent catastrophic forgetting in continual learning with novel hybrid classical-quantum neural networks.
We also claim that if the model is trained with these explanations, it tends to give better performance and learn specific features that are far from the decision boundary.
arXiv Detail & Related papers (2023-05-16T18:19:12Z) - On the balance between the training time and interpretability of neural
ODE for time series modelling [77.34726150561087]
The paper shows that modern neural ODE cannot be reduced to simpler models for time-series modelling applications.
The complexity of neural ODE is compared to or exceeds the conventional time-series modelling tools.
We propose a new view on time-series modelling using combined neural networks and an ODE system approach.
arXiv Detail & Related papers (2022-06-07T13:49:40Z) - Rethinking Nearest Neighbors for Visual Classification [56.00783095670361]
k-NN is a lazy learning method that aggregates the distance between the test image and top-k neighbors in a training set.
We adopt k-NN with pre-trained visual representations produced by either supervised or self-supervised methods in two steps.
Via extensive experiments on a wide range of classification tasks, our study reveals the generality and flexibility of k-NN integration.
arXiv Detail & Related papers (2021-12-15T20:15:01Z) - Learning Adaptive Embedding Considering Incremental Class [55.21855842960139]
Class-Incremental Learning (CIL) aims to train a reliable model with the streaming data, which emerges unknown classes sequentially.
Different from traditional closed set learning, CIL has two main challenges: 1) Novel class detection.
After the novel classes are detected, the model needs to be updated without re-training using entire previous data.
arXiv Detail & Related papers (2020-08-31T04:11:24Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z) - The large learning rate phase of deep learning: the catapult mechanism [50.23041928811575]
We present a class of neural networks with solvable training dynamics.
We find good agreement between our model's predictions and training dynamics in realistic deep learning settings.
We believe our results shed light on characteristics of models trained at different learning rates.
arXiv Detail & Related papers (2020-03-04T17:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.