Active Learning in Symbolic Regression with Physical Constraints
- URL: http://arxiv.org/abs/2305.10379v3
- Date: Fri, 9 Aug 2024 21:06:15 GMT
- Title: Active Learning in Symbolic Regression with Physical Constraints
- Authors: Jorge Medina, Andrew D. White,
- Abstract summary: Evolutionary symbolic regression (SR) fits a symbolic equation to data, which gives a concise interpretable model.
We explore using SR as a method to propose which data to gather in an active learning setting with physical constraints.
- Score: 0.4037357056611557
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Evolutionary symbolic regression (SR) fits a symbolic equation to data, which gives a concise interpretable model. We explore using SR as a method to propose which data to gather in an active learning setting with physical constraints. SR with active learning proposes which experiments to do next. Active learning is done with query by committee, where the Pareto frontier of equations is the committee. The physical constraints improve proposed equations in very low data settings. These approaches reduce the data required for SR and achieves state of the art results in data required to rediscover known equations.
Related papers
- Ab initio nonparametric variable selection for scalable Symbolic Regression with large $p$ [2.222138965069487]
Symbolic regression (SR) is a powerful technique for discovering symbolic expressions that characterize nonlinear relationships in data.
Existing SR methods do not scale to datasets with a large number of input variables, which are common in modern scientific applications.
We propose PAN+SR, which combines ab initio nonparametric variable selection with SR to efficiently pre-screen large input spaces.
arXiv Detail & Related papers (2024-10-17T15:41:06Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - A Transformer Model for Symbolic Regression towards Scientific Discovery [11.827358526480323]
Symbolic Regression (SR) searches for mathematical expressions which best describe numerical datasets.
We propose a new Transformer model aiming at Symbolic Regression particularly focused on its application for Scientific Discovery.
We apply our best model to the SRSD datasets which yields state-of-the-art results using the normalized tree-based edit distance.
arXiv Detail & Related papers (2023-12-07T06:27:48Z) - Discovering Interpretable Physical Models using Symbolic Regression and
Discrete Exterior Calculus [55.2480439325792]
We propose a framework that combines Symbolic Regression (SR) and Discrete Exterior Calculus (DEC) for the automated discovery of physical models.
DEC provides building blocks for the discrete analogue of field theories, which are beyond the state-of-the-art applications of SR to physical problems.
We prove the effectiveness of our methodology by re-discovering three models of Continuum Physics from synthetic experimental data.
arXiv Detail & Related papers (2023-10-10T13:23:05Z) - Incorporating Background Knowledge in Symbolic Regression using a
Computer Algebra System [0.0]
Symbolic Regression (SR) can generate interpretable, concise expressions that fit a given dataset.
We specifically examine the addition of constraints to traditional genetic algorithm (GA) based SR (PySR) as well as a Markov-chain Monte Carlo (MCMC) based Bayesian SR architecture.
arXiv Detail & Related papers (2023-01-27T18:59:25Z) - Rethinking Symbolic Regression Datasets and Benchmarks for Scientific
Discovery [12.496525234064888]
This paper revisits datasets and evaluation criteria for Symbolic Regression (SR)
We recreate 120 datasets to discuss the performance of symbolic regression for scientific discovery (SRSD)
arXiv Detail & Related papers (2022-06-21T17:15:45Z) - `Next Generation' Reservoir Computing: an Empirical Data-Driven
Expression of Dynamical Equations in Time-Stepping Form [0.0]
Next generation reservoir computing based on nonlinear vector autoregression is applied to emulate simple dynamical system models.
It is also shown that the approach can be extended to produce high-order numerical schemes directly from data.
The impacts of the presence of noise and temporal sparsity in the training set is examined to gauge the potential use of this method for more realistic applications.
arXiv Detail & Related papers (2022-01-13T20:13:33Z) - Neural Symbolic Regression that Scales [58.45115548924735]
We introduce the first symbolic regression method that leverages large scale pre-training.
We procedurally generate an unbounded set of equations, and simultaneously pre-train a Transformer to predict the symbolic equation from a corresponding set of input-output-pairs.
arXiv Detail & Related papers (2021-06-11T14:35:22Z) - A Hypergradient Approach to Robust Regression without Correspondence [85.49775273716503]
We consider a variant of regression problem, where the correspondence between input and output data is not available.
Most existing methods are only applicable when the sample size is small.
We propose a new computational framework -- ROBOT -- for the shuffled regression problem.
arXiv Detail & Related papers (2020-11-30T21:47:38Z) - Exposing Shallow Heuristics of Relation Extraction Models with Challenge
Data [49.378860065474875]
We identify failure modes of SOTA relation extraction (RE) models trained on TACRED.
By adding some of the challenge data as training examples, the performance of the model improves.
arXiv Detail & Related papers (2020-10-07T21:17:25Z) - Closed Loop Neural-Symbolic Learning via Integrating Neural Perception,
Grammar Parsing, and Symbolic Reasoning [134.77207192945053]
Prior methods learn the neural-symbolic models using reinforcement learning approaches.
We introduce the textbfgrammar model as a textitsymbolic prior to bridge neural perception and symbolic reasoning.
We propose a novel textbfback-search algorithm which mimics the top-down human-like learning procedure to propagate the error.
arXiv Detail & Related papers (2020-06-11T17:42:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.