Sampling from Pre-Images to Learn Heuristic Functions for Classical
Planning
- URL: http://arxiv.org/abs/2207.03336v1
- Date: Thu, 7 Jul 2022 14:42:31 GMT
- Title: Sampling from Pre-Images to Learn Heuristic Functions for Classical
Planning
- Authors: Stefan O'Toole, Miquel Ramirez, Nir Lipovetzky, Adrian R. Pearce
- Abstract summary: We introduce a new algorithm, Regression based Supervised Learning (RSL), for learning per instance Neural Network (NN) defined functions for classical planning problems.
RSL outperforms, in terms of coverage, previous classical planning NNs functions while requiring two orders of magnitude less training time.
- Score: 8.000374471991247
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a new algorithm, Regression based Supervised Learning (RSL), for
learning per instance Neural Network (NN) defined heuristic functions for
classical planning problems. RSL uses regression to select relevant sets of
states at a range of different distances from the goal. RSL then formulates a
Supervised Learning problem to obtain the parameters that define the NN
heuristic, using the selected states labeled with exact or estimated distances
to goal states. Our experimental study shows that RSL outperforms, in terms of
coverage, previous classical planning NN heuristics functions while requiring
two orders of magnitude less training time.
Related papers
- Using predefined vector systems as latent space configuration for neural network supervised training on data with arbitrarily large number of classes [0.0]
Supervised learning (SL) methods are indispensable for neural network (NN) training used to perform classification tasks.<n>We propose a methodology that allows one to train the same NN architecture regardless of the number of classes.
arXiv Detail & Related papers (2025-10-05T08:28:37Z) - Online Decision-Focused Learning [63.83903681295497]
Decision-focused learning (DFL) is an increasingly popular paradigm for training predictive models whose outputs are used in decision-making tasks.<n>We investigate DFL in dynamic environments where the objective function does not evolve over time.<n>We establish bounds on the expected dynamic regret, both when decision space is a simplex and when it is a general bounded convex polytope.
arXiv Detail & Related papers (2025-05-19T10:40:30Z) - Unveiling the Power of Sparse Neural Networks for Feature Selection [60.50319755984697]
Sparse Neural Networks (SNNs) have emerged as powerful tools for efficient feature selection.
We show that SNNs trained with dynamic sparse training (DST) algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
Our findings show that feature selection with SNNs trained with DST algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
arXiv Detail & Related papers (2024-08-08T16:48:33Z) - Planning in a recurrent neural network that plays Sokoban [6.059513516334741]
How a neural network generalizes to novel situations depends on whether it has learned to select actionsally via a planning process.
We find that a recurrent NN trained to play Sokoban appears to plan, with extra computation steps improving the RNN's success rate.
We perform model surgery, enabling the convolutional NN to generalize beyond its 10x10 architectural limit.
arXiv Detail & Related papers (2024-07-22T06:57:34Z) - Continual Learning via Sequential Function-Space Variational Inference [65.96686740015902]
We propose an objective derived by formulating continual learning as sequential function-space variational inference.
Compared to objectives that directly regularize neural network predictions, the proposed objective allows for more flexible variational distributions.
We demonstrate that, across a range of task sequences, neural networks trained via sequential function-space variational inference achieve better predictive accuracy than networks trained with related methods.
arXiv Detail & Related papers (2023-12-28T18:44:32Z) - Re-visiting Reservoir Computing architectures optimized by Evolutionary
Algorithms [0.0]
Evolutionary Algorithms (EAs) have been applied to improve Neural Networks (NNs) architectures.
We provide a systematic brief survey about applications of EAs on the specific domain of the recurrent NNs named Reservoir Computing (RC)
arXiv Detail & Related papers (2022-11-11T14:50:54Z) - Reinforcement Learning for Classical Planning: Viewing Heuristics as
Dense Reward Generators [54.6441336539206]
We propose to leverage domain-independent functions commonly used in the classical planning literature to improve the sample efficiency of RL.
These classicals act as dense reward generators to alleviate the sparse-rewards issue and enable our RL agent to learn domain-specific value functions as residuals.
We demonstrate on several classical planning domains that using classical logics for RL allows for good sample efficiency compared to sparse-reward RL.
arXiv Detail & Related papers (2021-09-30T03:36:01Z) - Learning Regularization Parameters of Inverse Problems via Deep Neural
Networks [0.0]
We consider a supervised learning approach, where a network is trained to approximate the mapping from observation data to regularization parameters.
We show that a wide variety of regularization functionals, forward models, and noise models may be considered.
The network-obtained regularization parameters can be computed more efficiently and may even lead to more accurate solutions.
arXiv Detail & Related papers (2021-04-14T02:38:38Z) - NSL: Hybrid Interpretable Learning From Noisy Raw Data [66.15862011405882]
This paper introduces a hybrid neural-symbolic learning framework, called NSL, that learns interpretable rules from labelled unstructured data.
NSL combines pre-trained neural networks for feature extraction with FastLAS, a state-of-the-art ILP system for rule learning under the answer set semantics.
We demonstrate that NSL is able to learn robust rules from MNIST data and achieve comparable or superior accuracy when compared to neural network and random forest baselines.
arXiv Detail & Related papers (2020-12-09T13:02:44Z) - Spiking Neural Networks -- Part II: Detecting Spatio-Temporal Patterns [38.518936229794214]
Spiking Neural Networks (SNNs) have the unique ability to detect information in encoded-temporal signals.
We review models and training algorithms for the dominant approach that considers SNNs as a Recurrent Neural Network (RNN)
We describe an alternative approach that relies on probabilistic models for spiking neurons, allowing the derivation of local learning rules via gradient estimates.
arXiv Detail & Related papers (2020-10-27T11:47:42Z) - A Novel Neural Network Training Framework with Data Assimilation [2.948167339160823]
A gradient-free training framework based on data assimilation is proposed to avoid the calculation of gradients.
The results show that the proposed training framework performed better than the gradient decent method.
arXiv Detail & Related papers (2020-10-06T11:12:23Z) - Persistent Neurons [4.061135251278187]
We propose a trajectory-based strategy that optimize the learning task using information from previous solutions.
Persistent neurons can be regarded as a method with gradient informed bias where individual updates are corrupted by deterministic error terms.
We evaluate the full and partial persistent model and show it can be used to boost the performance on a range of NN structures.
arXiv Detail & Related papers (2020-07-02T22:36:49Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - AdaS: Adaptive Scheduling of Stochastic Gradients [50.80697760166045]
We introduce the notions of textit"knowledge gain" and textit"mapping condition" and propose a new algorithm called Adaptive Scheduling (AdaS)
Experimentation reveals that, using the derived metrics, AdaS exhibits: (a) faster convergence and superior generalization over existing adaptive learning methods; and (b) lack of dependence on a validation set to determine when to stop training.
arXiv Detail & Related papers (2020-06-11T16:36:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.