Sym-Q: Adaptive Symbolic Regression via Sequential Decision-Making
- URL: http://arxiv.org/abs/2402.05306v1
- Date: Wed, 7 Feb 2024 22:53:54 GMT
- Title: Sym-Q: Adaptive Symbolic Regression via Sequential Decision-Making
- Authors: Yuan Tian, Wenqi Zhou, Hao Dong, David S. Kammer, Olga Fink
- Abstract summary: Symbolic regression holds great potential for uncovering underlying mathematical and physical relationships from empirical data.
Existing transformer-based models face challenges in terms of generalizability and adaptability.
We introduce Symbolic Q-network (Sym-Q), a novel reinforcement learning-based model that redefines symbolic regression as a sequential decision-making task.
- Score: 13.419259918160321
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Symbolic regression holds great potential for uncovering underlying
mathematical and physical relationships from empirical data. While existing
transformer-based models have recently achieved significant success in this
domain, they face challenges in terms of generalizability and adaptability.
Typically, in cases where the output expressions do not adequately fit
experimental data, the models lack efficient mechanisms to adapt or modify the
expression. This inflexibility hinders their application in real-world
scenarios, particularly in discovering unknown physical or biological
relationships. Inspired by how human experts refine and adapt expressions, we
introduce Symbolic Q-network (Sym-Q), a novel reinforcement learning-based
model that redefines symbolic regression as a sequential decision-making task.
Sym-Q leverages supervised demonstrations and refines expressions based on
reward signals indicating the quality of fitting precision. Its distinctive
ability to manage the complexity of expression trees and perform precise
step-wise updates significantly enhances flexibility and efficiency. Our
results demonstrate that Sym-Q excels not only in recovering underlying
mathematical structures but also uniquely learns to efficiently refine the
output expression based on reward signals, thereby discovering underlying
expressions. Sym-Q paves the way for more intuitive and impactful discoveries
in physical science, marking a substantial advancement in the field of symbolic
regression.
Related papers
- Interactive Symbolic Regression through Offline Reinforcement Learning: A Co-Design Framework [11.804368618793273]
Symbolic Regression holds great potential for uncovering underlying mathematical and physical relationships from observed data.
Current state-of-the-art approaches typically do not consider the integration of domain experts' prior knowledge.
We propose the Symbolic Q-network (Sym-Q), an advanced interactive framework for large-scale symbolic regression.
arXiv Detail & Related papers (2025-02-05T06:26:49Z) - Test-time regression: a unifying framework for designing sequence models with associative memory [24.915262407519876]
We show that effective sequence models must be able to perform associative recall.
Our key insight is that memorizing input tokens through an associative memory is equivalent to performing regression at test-time.
We show numerous recent architectures -- including linear attention models, their gated variants, state-space models, online learners, and softmax attention -- emerge naturally as specific approaches to test-time regression.
arXiv Detail & Related papers (2025-01-21T18:32:31Z) - Representation Learning with Parameterised Quantum Circuits for Advancing Speech Emotion Recognition [37.98283871637917]
Speech Emotion Recognition (SER) is a complex task in human-computer interaction due to the intricate dependencies of features and the overlapping nature of emotional expressions conveyed through speech.
This paper introduces a hybrid classical-quantum framework that integrates volutionised Quantum Circuits with conventional Conal Neural Network (CNN) architectures.
By leveraging quantum properties such as superposition and entanglement, the proposed model enhances feature representation and captures complex dependencies more effectively than classical methods.
arXiv Detail & Related papers (2025-01-21T11:23:38Z) - Entropy-Regularized Token-Level Policy Optimization for Language Agent Reinforcement [67.1393112206885]
Large Language Models (LLMs) have shown promise as intelligent agents in interactive decision-making tasks.
We introduce Entropy-Regularized Token-level Policy Optimization (ETPO), an entropy-augmented RL method tailored for optimizing LLMs at the token level.
We assess the effectiveness of ETPO within a simulated environment that models data science code generation as a series of multi-step interactive tasks.
arXiv Detail & Related papers (2024-02-09T07:45:26Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - Pointer Networks with Q-Learning for Combinatorial Optimization [55.2480439325792]
We introduce the Pointer Q-Network (PQN), a hybrid neural architecture that integrates model-free Q-value policy approximation with Pointer Networks (Ptr-Nets)
Our empirical results demonstrate the efficacy of this approach, also testing the model in unstable environments.
arXiv Detail & Related papers (2023-11-05T12:03:58Z) - Understanding Augmentation-based Self-Supervised Representation Learning
via RKHS Approximation and Regression [53.15502562048627]
Recent work has built the connection between self-supervised learning and the approximation of the top eigenspace of a graph Laplacian operator.
This work delves into a statistical analysis of augmentation-based pretraining.
arXiv Detail & Related papers (2023-06-01T15:18:55Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Transformer-based Planning for Symbolic Regression [18.90700817248397]
We propose TPSR, a Transformer-based Planning strategy for Symbolic Regression.
Unlike conventional decoding strategies, TPSR enables the integration of non-differentiable feedback, such as fitting accuracy and complexity.
Our approach outperforms state-of-the-art methods, enhancing the model's fitting-complexity trade-off, Symbolic abilities, and robustness to noise.
arXiv Detail & Related papers (2023-03-13T03:29:58Z) - IQ-Learn: Inverse soft-Q Learning for Imitation [95.06031307730245]
imitation learning from a small amount of expert data can be challenging in high-dimensional environments with complex dynamics.
Behavioral cloning is a simple method that is widely used due to its simplicity of implementation and stable convergence.
We introduce a method for dynamics-aware IL which avoids adversarial training by learning a single Q-function.
arXiv Detail & Related papers (2021-06-23T03:43:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.