One-class systems seamlessly fit in the forward-forward algorithm
- URL: http://arxiv.org/abs/2306.15188v1
- Date: Tue, 27 Jun 2023 04:14:03 GMT
- Title: One-class systems seamlessly fit in the forward-forward algorithm
- Authors: Michael Hopwood
- Abstract summary: We present a new method of training neural networks by updating weights during an inference, performing parameter updates for each layer individually.
This immediately reduces memory requirements during training and may lead to many more benefits, like seamless online training.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The forward-forward algorithm presents a new method of training neural
networks by updating weights during an inference, performing parameter updates
for each layer individually. This immediately reduces memory requirements
during training and may lead to many more benefits, like seamless online
training. This method relies on a loss ("goodness") function that can be
evaluated on the activations of each layer, of which can have a varied
parameter size, depending on the hyperparamaterization of the network. In the
seminal paper, a goodness function was proposed to fill this need; however, if
placed in a one-class problem context, one need not pioneer a new loss because
these functions can innately handle dynamic network sizes. In this paper, we
investigate the performance of deep one-class objective functions when trained
in a forward-forward fashion. The code is available at
\url{https://github.com/MichaelHopwood/ForwardForwardOneclass}.
Related papers
- InRank: Incremental Low-Rank Learning [85.6380047359139]
gradient-based training implicitly regularizes neural networks towards low-rank solutions through a gradual increase of the rank during training.
Existing training algorithms do not exploit the low-rank property to improve computational efficiency.
We design a new training algorithm Incremental Low-Rank Learning (InRank), which explicitly expresses cumulative weight updates as low-rank matrices.
arXiv Detail & Related papers (2023-06-20T03:03:04Z) - Slimmable Networks for Contrastive Self-supervised Learning [69.9454691873866]
Self-supervised learning makes significant progress in pre-training large models, but struggles with small models.
We introduce another one-stage solution to obtain pre-trained small models without the need for extra teachers.
A slimmable network consists of a full network and several weight-sharing sub-networks, which can be pre-trained once to obtain various networks.
arXiv Detail & Related papers (2022-09-30T15:15:05Z) - Training Your Sparse Neural Network Better with Any Mask [106.134361318518]
Pruning large neural networks to create high-quality, independently trainable sparse masks is desirable.
In this paper we demonstrate an alternative opportunity: one can customize the sparse training techniques to deviate from the default dense network training protocols.
Our new sparse training recipe is generally applicable to improving training from scratch with various sparse masks.
arXiv Detail & Related papers (2022-06-26T00:37:33Z) - Dimensionality Reduced Training by Pruning and Freezing Parts of a Deep
Neural Network, a Survey [69.3939291118954]
State-of-the-art deep learning models have a parameter count that reaches into the billions. Training, storing and transferring such models is energy and time consuming, thus costly.
Model compression lowers storage and transfer costs, and can further make training more efficient by decreasing the number of computations in the forward and/or backward pass.
This work is a survey on methods which reduce the number of trained weights in deep learning models throughout the training.
arXiv Detail & Related papers (2022-05-17T05:37:08Z) - Training Neural Networks with Fixed Sparse Masks [19.58969772430058]
Recent work has shown that it is possible to update only a small subset of the model's parameters during training.
We show that it is possible to induce a fixed sparse mask on the model's parameters that selects a subset to update over many iterations.
arXiv Detail & Related papers (2021-11-18T18:06:01Z) - Using UNet and PSPNet to explore the reusability principle of CNN
parameters [5.623232537411766]
Reusability of parameters in each layer of a deep convolutional neural network is experimentally quantified.
Running mean and running variance plays an important role than Weight and Bias in BN layer.
The bias in Convolution layers are not sensitive, and it can be reused directly.
arXiv Detail & Related papers (2020-08-08T01:51:08Z) - Semantic Drift Compensation for Class-Incremental Learning [48.749630494026086]
Class-incremental learning of deep networks sequentially increases the number of classes to be classified.
We propose a new method to estimate the drift, called semantic drift, of features and compensate for it without the need of any exemplars.
arXiv Detail & Related papers (2020-04-01T13:31:19Z) - Cooperative Initialization based Deep Neural Network Training [35.14235994478142]
Our approach uses multiple activation functions in the initial few epochs for the update of all sets of weight parameters while training the network.
Our approach outperforms various baselines and, at the same time, performs well over various tasks such as classification and detection.
arXiv Detail & Related papers (2020-01-05T14:08:46Z) - Side-Tuning: A Baseline for Network Adaptation via Additive Side
Networks [95.51368472949308]
Adaptation can be useful in cases when training data is scarce, or when one wishes to encode priors in the network.
In this paper, we propose a straightforward alternative: side-tuning.
arXiv Detail & Related papers (2019-12-31T18:52:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.