RankNEAT: Outperforming Stochastic Gradient Search in Preference
Learning Tasks
- URL: http://arxiv.org/abs/2204.06901v1
- Date: Thu, 14 Apr 2022 12:01:00 GMT
- Title: RankNEAT: Outperforming Stochastic Gradient Search in Preference
Learning Tasks
- Authors: Kosmas Pinitas, Konstantinos Makantasis, Antonios Liapis, Georgios N.
Yannakakis
- Abstract summary: gradient descent (SGD) is a premium optimization method for training neural networks.
We introduce the RankNEAT algorithm which learns to rank through neuroevolution of augmenting topologies.
Results suggest that RankNEAT is a viable and highly efficient evolutionary alternative to preference learning.
- Score: 2.570570340104555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic gradient descent (SGD) is a premium optimization method for
training neural networks, especially for learning objectively defined labels
such as image objects and events. When a neural network is instead faced with
subjectively defined labels--such as human demonstrations or annotations--SGD
may struggle to explore the deceptive and noisy loss landscapes caused by the
inherent bias and subjectivity of humans. While neural networks are often
trained via preference learning algorithms in an effort to eliminate such data
noise, the de facto training methods rely on gradient descent. Motivated by the
lack of empirical studies on the impact of evolutionary search to the training
of preference learners, we introduce the RankNEAT algorithm which learns to
rank through neuroevolution of augmenting topologies. We test the hypothesis
that RankNEAT outperforms traditional gradient-based preference learning within
the affective computing domain, in particular predicting annotated player
arousal from the game footage of three dissimilar games. RankNEAT yields
superior performances compared to the gradient-based preference learner
(RankNet) in the majority of experiments since its architecture optimization
capacity acts as an efficient feature selection mechanism, thereby, eliminating
overfitting. Results suggest that RankNEAT is a viable and highly efficient
evolutionary alternative to preference learning.
Related papers
- Simmering: Sufficient is better than optimal for training neural networks [0.0]
We introduce simmering, a physics-based method that trains neural networks to generate weights and biases that are merely good enough''
We show that simmering corrects neural networks that are overfit by Adam, and show that simmering avoids overfitting if deployed from the outset.
Our results question optimization as a paradigm for neural network training, and leverage information-geometric arguments to point to the existence of classes of sufficient training algorithms.
arXiv Detail & Related papers (2024-10-25T18:02:08Z) - Newton Losses: Using Curvature Information for Learning with Differentiable Algorithms [80.37846867546517]
We show how to train eight different neural networks with custom objectives.
We exploit their second-order information via their empirical Fisherssian matrices.
We apply Loss Lossiable algorithms to achieve significant improvements for less differentiable algorithms.
arXiv Detail & Related papers (2024-10-24T18:02:11Z) - Forward-Forward Algorithm for Hyperspectral Image Classification: A
Preliminary Study [0.0]
Forward-forward algorithm (FFA) computes local goodness functions to optimize network parameters.
This study investigates the application of FFA for hyperspectral image classification.
arXiv Detail & Related papers (2023-07-01T05:39:28Z) - Lottery Tickets in Evolutionary Optimization: On Sparse
Backpropagation-Free Trainability [0.0]
We study gradient descent (GD)-based sparse training and evolution strategies (ES)
We find that ES explore diverse and flat local optima and do not preserve linear mode connectivity across sparsity levels and independent runs.
arXiv Detail & Related papers (2023-05-31T15:58:54Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Neuroevolution of Physics-Informed Neural Nets: Benchmark Problems and
Comparative Results [25.12291688711645]
Physics-informed neural networks (PINNs) are one of the key techniques at the forefront of recent advances.
PINNs' unique loss formulations lead to a high degree of complexity and ruggedness that may not be conducive for gradient descent.
Neuroevolution algorithms, with their superior global search capacity, may be a better choice for PINNs.
arXiv Detail & Related papers (2022-12-15T05:54:16Z) - Towards Theoretically Inspired Neural Initialization Optimization [66.04735385415427]
We propose a differentiable quantity, named GradCosine, with theoretical insights to evaluate the initial state of a neural network.
We show that both the training and test performance of a network can be improved by maximizing GradCosine under norm constraint.
Generalized from the sample-wise analysis into the real batch setting, NIO is able to automatically look for a better initialization with negligible cost.
arXiv Detail & Related papers (2022-10-12T06:49:16Z) - Scaling Forward Gradient With Local Losses [117.22685584919756]
Forward learning is a biologically plausible alternative to backprop for learning deep neural networks.
We show that it is possible to substantially reduce the variance of the forward gradient by applying perturbations to activations rather than weights.
Our approach matches backprop on MNIST and CIFAR-10 and significantly outperforms previously proposed backprop-free algorithms on ImageNet.
arXiv Detail & Related papers (2022-10-07T03:52:27Z) - Improving Music Performance Assessment with Contrastive Learning [78.8942067357231]
This study investigates contrastive learning as a potential method to improve existing MPA systems.
We introduce a weighted contrastive loss suitable for regression tasks applied to a convolutional neural network.
Our results show that contrastive-based methods are able to match and exceed SoTA performance for MPA regression tasks.
arXiv Detail & Related papers (2021-08-03T19:24:25Z) - Analytically Tractable Inference in Deep Neural Networks [0.0]
Tractable Approximate Inference (TAGI) algorithm was shown to be a viable and scalable alternative to backpropagation for shallow fully-connected neural networks.
We are demonstrating how TAGI matches or exceeds the performance of backpropagation, for training classic deep neural network architectures.
arXiv Detail & Related papers (2021-03-09T14:51:34Z) - The Break-Even Point on Optimization Trajectories of Deep Neural
Networks [64.7563588124004]
We argue for the existence of the "break-even" point on this trajectory.
We show that using a large learning rate in the initial phase of training reduces the variance of the gradient.
We also show that using a low learning rate results in bad conditioning of the loss surface even for a neural network with batch normalization layers.
arXiv Detail & Related papers (2020-02-21T22:55:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.