Toward Physically Plausible Data-Driven Models: A Novel Neural Network
Approach to Symbolic Regression
- URL: http://arxiv.org/abs/2302.00773v3
- Date: Tue, 27 Jun 2023 12:36:32 GMT
- Title: Toward Physically Plausible Data-Driven Models: A Novel Neural Network
Approach to Symbolic Regression
- Authors: Ji\v{r}\'i Kubal\'ik, Erik Derner, Robert Babu\v{s}ka
- Abstract summary: This paper proposes a novel neural network-based symbolic regression method.
It constructs physically plausible models based on even very small training data sets and prior knowledge about the system.
We experimentally evaluate the approach on four test systems: the TurtleBot 2 mobile robot, the magnetic manipulation system, the equivalent resistance of two resistors in parallel, and the longitudinal force of the anti-lock braking system.
- Score: 2.7071541526963805
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many real-world systems can be described by mathematical models that are
human-comprehensible, easy to analyze and help explain the system's behavior.
Symbolic regression is a method that can automatically generate such models
from data. Historically, symbolic regression has been predominantly realized by
genetic programming, a method that evolves populations of candidate solutions
that are subsequently modified by genetic operators crossover and mutation.
However, this approach suffers from several deficiencies: it does not scale
well with the number of variables and samples in the training data - models
tend to grow in size and complexity without an adequate accuracy gain, and it
is hard to fine-tune the model coefficients using just genetic operators.
Recently, neural networks have been applied to learn the whole analytic model,
i.e., its structure and the coefficients, using gradient-based optimization
algorithms. This paper proposes a novel neural network-based symbolic
regression method that constructs physically plausible models based on even
very small training data sets and prior knowledge about the system. The method
employs an adaptive weighting scheme to effectively deal with multiple loss
function terms and an epoch-wise learning process to reduce the chance of
getting stuck in poor local optima. Furthermore, we propose a parameter-free
method for choosing the model with the best interpolation and extrapolation
performance out of all the models generated throughout the whole learning
process. We experimentally evaluate the approach on four test systems: the
TurtleBot 2 mobile robot, the magnetic manipulation system, the equivalent
resistance of two resistors in parallel, and the longitudinal force of the
anti-lock braking system. The results clearly show the potential of the method
to find parsimonious models that comply with the prior knowledge provided.
Related papers
- Exploring hyperelastic material model discovery for human brain cortex:
multivariate analysis vs. artificial neural network approaches [10.003764827561238]
This study aims to identify the most favorable material model for human brain tissue.
We apply artificial neural network and multiple regression methods to a generalization of widely accepted classic models.
arXiv Detail & Related papers (2023-10-16T18:49:59Z) - Autoregressive models for biomedical signal processing [0.0]
We present a framework for autoregressive modelling that incorporates uncertainties explicitly via a loss function.
Our work shows that the procedure is able to successfully denoise time series and successfully reconstruct system parameters.
This new paradigm can be used in a multitude of applications in neuroscience such as brain-computer interface data analysis and better understanding of brain dynamics in diseases such as epilepsy.
arXiv Detail & Related papers (2023-04-17T08:57:36Z) - Online model error correction with neural networks in the incremental
4D-Var framework [0.0]
We develop a new weak-constraint 4D-Var formulation which can be used to train a neural network for online model error correction.
The method is implemented in the ECMWF Object-Oriented Prediction System.
The results confirm that online learning is effective and yields a more accurate model error correction than offline learning.
arXiv Detail & Related papers (2022-10-25T07:45:33Z) - Supervised Parameter Estimation of Neuron Populations from Multiple
Firing Events [3.2826301276626273]
We study an automatic approach of learning the parameters of neuron populations from a training set consisting of pairs of spiking series and parameter labels via supervised learning.
We simulate many neuronal populations at computation at different parameter settings using a neuron model.
We then compare their performance against classical approaches including a genetic search, Bayesian sequential estimation, and a random walk approximate model.
arXiv Detail & Related papers (2022-10-02T03:17:05Z) - Neural parameter calibration for large-scale multi-agent models [0.7734726150561089]
We present a method to retrieve accurate probability densities for parameters using neural equations.
The two combined create a powerful tool that can quickly estimate densities on model parameters, even for very large systems.
arXiv Detail & Related papers (2022-09-27T17:36:26Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Using Data Assimilation to Train a Hybrid Forecast System that Combines
Machine-Learning and Knowledge-Based Components [52.77024349608834]
We consider the problem of data-assisted forecasting of chaotic dynamical systems when the available data is noisy partial measurements.
We show that by using partial measurements of the state of the dynamical system, we can train a machine learning model to improve predictions made by an imperfect knowledge-based model.
arXiv Detail & Related papers (2021-02-15T19:56:48Z) - Large-scale Neural Solvers for Partial Differential Equations [48.7576911714538]
Solving partial differential equations (PDE) is an indispensable part of many branches of science as many processes can be modelled in terms of PDEs.
Recent numerical solvers require manual discretization of the underlying equation as well as sophisticated, tailored code for distributed computing.
We examine the applicability of continuous, mesh-free neural solvers for partial differential equations, physics-informed neural networks (PINNs)
We discuss the accuracy of GatedPINN with respect to analytical solutions -- as well as state-of-the-art numerical solvers, such as spectral solvers.
arXiv Detail & Related papers (2020-09-08T13:26:51Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.