SymbolNet: Neural Symbolic Regression with Adaptive Dynamic Pruning
- URL: http://arxiv.org/abs/2401.09949v1
- Date: Thu, 18 Jan 2024 12:51:38 GMT
- Title: SymbolNet: Neural Symbolic Regression with Adaptive Dynamic Pruning
- Authors: Ho Fung Tsoi, Vladimir Loncar, Sridhara Dasu, Philip Harris
- Abstract summary: We propose a neural network approach to symbolic regression in a novel framework that enables dynamic pruning of model weights, input features, and mathematical operators in a single training.
In contrast to most existing symbolic regression methods that cannot efficiently handle datasets with more than $O$ inputs, we demonstrate the effectiveness of our model on the LHC jet tagging task.
- Score: 1.1313585906598267
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Contrary to the use of genetic programming, the neural network approach to
symbolic regression can scale well with high input dimension and leverage
gradient methods for faster equation searching. Common ways of constraining
expression complexity have relied on multistage pruning methods with
fine-tuning, but these often lead to significant performance loss. In this
work, we propose SymbolNet, a neural network approach to symbolic regression in
a novel framework that enables dynamic pruning of model weights, input
features, and mathematical operators in a single training, where both training
loss and expression complexity are optimized simultaneously. We introduce a
sparsity regularization term per pruning type, which can adaptively adjust its
own strength and lead to convergence to a target sparsity level. In contrast to
most existing symbolic regression methods that cannot efficiently handle
datasets with more than $O$(10) inputs, we demonstrate the effectiveness of our
model on the LHC jet tagging task (16 inputs), MNIST (784 inputs), and SVHN
(3072 inputs).
Related papers
- An Efficient Approach to Regression Problems with Tensor Neural Networks [5.345144592056051]
This paper introduces a tensor neural network (TNN) to address nonparametric regression problems.
The TNN demonstrates superior performance compared to conventional Feed-Forward Networks (FFN) and Radial Basis Function Networks (RBN)
A significant innovation in our approach is the integration of statistical regression and numerical integration within the TNN framework.
arXiv Detail & Related papers (2024-06-14T03:38:40Z) - Distributed Representations Enable Robust Multi-Timescale Symbolic Computation in Neuromorphic Hardware [3.961418890143814]
We describe a single-shot weight learning scheme to embed robust multi-timescale dynamics into attractor-based RSNNs.
We embed finite state machines into the RSNN dynamics by superimposing a symmetric autoassociative weight matrix.
This work introduces a scalable approach to embed robust symbolic computation through recurrent dynamics into neuromorphic hardware.
arXiv Detail & Related papers (2024-05-02T14:11:50Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - RWKV: Reinventing RNNs for the Transformer Era [54.716108899349614]
We propose a novel model architecture that combines the efficient parallelizable training of transformers with the efficient inference of RNNs.
We scale our models as large as 14 billion parameters, by far the largest dense RNN ever trained, and find RWKV performs on par with similarly sized Transformers.
arXiv Detail & Related papers (2023-05-22T13:57:41Z) - Return of the RNN: Residual Recurrent Networks for Invertible Sentence
Embeddings [0.0]
This study presents a novel model for invertible sentence embeddings using a residual recurrent network trained on an unsupervised encoding task.
Rather than the probabilistic outputs common to neural machine translation models, our approach employs a regression-based output layer to reconstruct the input sequence's word vectors.
The model achieves high accuracy and fast training with the ADAM, a significant finding given that RNNs typically require memory units, such as LSTMs, or second-order optimization methods.
arXiv Detail & Related papers (2023-03-23T15:59:06Z) - Transformer-based Planning for Symbolic Regression [18.90700817248397]
We propose TPSR, a Transformer-based Planning strategy for Symbolic Regression.
Unlike conventional decoding strategies, TPSR enables the integration of non-differentiable feedback, such as fitting accuracy and complexity.
Our approach outperforms state-of-the-art methods, enhancing the model's fitting-complexity trade-off, Symbolic abilities, and robustness to noise.
arXiv Detail & Related papers (2023-03-13T03:29:58Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Symplectically Integrated Symbolic Regression of Hamiltonian Dynamical
Systems [11.39873640706974]
Symplectically Integrated Symbolic Regression (SISR) is a novel technique for learning physical governing equations from data.
SISR employs a deep symbolic regression approach, using a multi-layer LSTM-RNN with mutation to probabilistically sample Hamiltonian symbolic expressions.
arXiv Detail & Related papers (2022-09-04T03:17:40Z) - Mitigating Performance Saturation in Neural Marked Point Processes:
Architectures and Loss Functions [50.674773358075015]
We propose a simple graph-based network structure called GCHP, which utilizes only graph convolutional layers.
We show that GCHP can significantly reduce training time and the likelihood ratio loss with interarrival time probability assumptions can greatly improve the model performance.
arXiv Detail & Related papers (2021-07-07T16:59:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.