Exploring Multi-view Symbolic Regression methods in physical sciences
- URL: http://arxiv.org/abs/2509.10500v1
- Date: Mon, 01 Sep 2025 12:43:34 GMT
- Title: Exploring Multi-view Symbolic Regression methods in physical sciences
- Authors: Etienne Russeil, Fabrício Olivetti de França, Konstantin Malanchev, Guillaume Moinard, Maxime Cherrey,
- Abstract summary: Multi-view Symbolic Regression (MvSR) searches for a parametric function capable of describing multiple datasets generated by the same phenomena.<n>We show that MvSR often achieve good accuracy while proposing solutions with only few free parameters.<n>We conclude by providing guidelines for future MvSR developments.
- Score: 0.7503129292751939
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Describing the world behavior through mathematical functions help scientists to achieve a better understanding of the inner mechanisms of different phenomena. Traditionally, this is done by deriving new equations from first principles and careful observations. A modern alternative is to automate part of this process with symbolic regression (SR). The SR algorithms search for a function that adequately fits the observed data while trying to enforce sparsity, in the hopes of generating an interpretable equation. A particularly interesting extension to these algorithms is the Multi-view Symbolic Regression (MvSR). It searches for a parametric function capable of describing multiple datasets generated by the same phenomena, which helps to mitigate the common problems of overfitting and data scarcity. Recently, multiple implementations added support to MvSR with small differences between them. In this paper, we test and compare MvSR as supported in Operon, PySR, phy-SO, and eggp, in different real-world datasets. We show that they all often achieve good accuracy while proposing solutions with only few free parameters. However, we find that certain features enable a more frequent generation of better models. We conclude by providing guidelines for future MvSR developments.
Related papers
- MMSR: Symbolic Regression is a Multimodal Task [12.660401635672967]
Symbolic regression was originally formulated as a optimization problem, and GP and reinforcement learning algorithms were used to solve it.
To solve this problem, researchers treat the mapping from data to expressions as a translation problem.
In this paper, we propose MMSR, which achieves the most advanced results on multiple mainstream datasets.
arXiv Detail & Related papers (2024-02-28T08:29:42Z) - Multi-View Symbolic Regression [1.2334534968968969]
We present Multi-View Symbolic Regression (MvSR), which takes into account multiple datasets simultaneously.
MvSR fits the evaluated expression to each independent dataset and returns a parametric family of functions.
We demonstrate the effectiveness of MvSR using data generated from known expressions, as well as real-world data from astronomy, chemistry and economy.
arXiv Detail & Related papers (2024-02-06T15:53:49Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - ParFam -- (Neural Guided) Symbolic Regression Based on Continuous Global Optimization [14.146976111782466]
We present our new approach ParFam to translate the discrete symbolic regression problem into a continuous one.<n>In combination with a global, this approach results in a highly effective method to tackle the problem of SR.<n>We also present an extension incorporating a pre-trained transformer network DL-ParFam to guide ParFam.
arXiv Detail & Related papers (2023-10-09T09:01:25Z) - Scalable Neural Symbolic Regression using Control Variables [7.725394912527969]
We propose ScaleSR, a scalable symbolic regression model that leverages control variables to enhance both accuracy and scalability.
The proposed method involves a four-step process. First, we learn a data generator from observed data using deep neural networks (DNNs)
Experimental results demonstrate that the proposed ScaleSR significantly outperforms state-of-the-art baselines in discovering mathematical expressions with multiple variables.
arXiv Detail & Related papers (2023-06-07T18:30:25Z) - Exhaustive Symbolic Regression [0.0]
Exhaustive Symbolic Regression (ESR) is a rigorous method for combining preferences into a single objective.
We apply it to a catalogue of cosmic chronometers and the Pantheon+ sample of supernovae to learn the Hubble rate.
We make our code and full equation sets publicly available.
arXiv Detail & Related papers (2022-11-21T13:48:52Z) - A Hypergradient Approach to Robust Regression without Correspondence [85.49775273716503]
We consider a variant of regression problem, where the correspondence between input and output data is not available.
Most existing methods are only applicable when the sample size is small.
We propose a new computational framework -- ROBOT -- for the shuffled regression problem.
arXiv Detail & Related papers (2020-11-30T21:47:38Z) - Learning outside the Black-Box: The pursuit of interpretable models [78.32475359554395]
This paper proposes an algorithm that produces a continuous global interpretation of any given continuous black-box function.
Our interpretation represents a leap forward from the previous state of the art.
arXiv Detail & Related papers (2020-11-17T12:39:44Z) - On Function Approximation in Reinforcement Learning: Optimism in the
Face of Large State Spaces [208.67848059021915]
We study the exploration-exploitation tradeoff at the core of reinforcement learning.
In particular, we prove that the complexity of the function class $mathcalF$ characterizes the complexity of the function.
Our regret bounds are independent of the number of episodes.
arXiv Detail & Related papers (2020-11-09T18:32:22Z) - Piecewise Linear Regression via a Difference of Convex Functions [50.89452535187813]
We present a new piecewise linear regression methodology that utilizes fitting a difference of convex functions (DC functions) to the data.
We empirically validate the method, showing it to be practically implementable, and to have comparable performance to existing regression/classification methods on real-world datasets.
arXiv Detail & Related papers (2020-07-05T18:58:47Z) - Symbolic Regression using Mixed-Integer Nonlinear Optimization [9.638685454900047]
The Symbolic Regression (SR) problem is a hard problem in machine learning.
We propose a hybrid algorithm that combines mixed-integer nonlinear optimization with explicit enumeration.
We show that our algorithm is competitive, for some synthetic data sets, with a state-of-the-art SR software and a recent physics-inspired method called AI Feynman.
arXiv Detail & Related papers (2020-06-11T20:53:17Z) - The data-driven physical-based equations discovery using evolutionary
approach [77.34726150561087]
We describe the algorithm for the mathematical equations discovery from the given observations data.
The algorithm combines genetic programming with the sparse regression.
It could be used for governing analytical equation discovery as well as for partial differential equations (PDE) discovery.
arXiv Detail & Related papers (2020-04-03T17:21:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.