A Deep Learning Framework for Multi-Operator Learning: Architectures and Approximation Theory
- URL: http://arxiv.org/abs/2510.25379v1
- Date: Wed, 29 Oct 2025 10:52:02 GMT
- Title: A Deep Learning Framework for Multi-Operator Learning: Architectures and Approximation Theory
- Authors: Adrien Weihs, Jingmin Sun, Zecheng Zhang, Hayden Schaeffer,
- Abstract summary: We study the problem of learning collections of operators and provide both theoretical and empirical advances.<n>We distinguish between two regimes: (i) multiple operator learning, where a single network represents a continuum of operators parameterized by a parametric function, and (ii) learning several distinct single operators, where each operator is learned independently.<n>Overall, this work establishes a unified theoretical and practical foundation for scalable operator learning across multiple operators.
- Score: 2.2731895181875346
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While many problems in machine learning focus on learning mappings between finite-dimensional spaces, scientific applications require approximating mappings between function spaces, i.e., operators. We study the problem of learning collections of operators and provide both theoretical and empirical advances. We distinguish between two regimes: (i) multiple operator learning, where a single network represents a continuum of operators parameterized by a parametric function, and (ii) learning several distinct single operators, where each operator is learned independently. For the multiple operator case, we introduce two new architectures, $\mathrm{MNO}$ and $\mathrm{MONet}$, and establish universal approximation results in three settings: continuous, integrable, or Lipschitz operators. For the latter, we further derive explicit scaling laws that quantify how the network size must grow to achieve a target approximation accuracy. For learning several single operators, we develop a framework for balancing architectural complexity across subnetworks and show how approximation order determines computational efficiency. Empirical experiments on parametric PDE benchmarks confirm the strong expressive power and efficiency of the proposed architectures. Overall, this work establishes a unified theoretical and practical foundation for scalable neural operator learning across multiple operators.
Related papers
- Max-Min Neural Network Operators For Approximation of Multivariate Functions [4.5657111459153885]
We develop a framework for approximation by max-min neural network operators.<n>We establish pointwise and uniform convergence theorems and derive estimates for the order of approximation.
arXiv Detail & Related papers (2026-01-12T06:14:05Z) - Fourier Neural Operators Explained: A Practical Perspective [75.12291469255794]
The Fourier Neural Operator (FNO) has become the most influential and widely adopted due to its elegant spectral formulation.<n>This guide aims to establish a clear and reliable framework for applying FNOs effectively across diverse scientific and engineering fields.
arXiv Detail & Related papers (2025-12-01T08:56:21Z) - Multi-Operator Few-Shot Learning for Generalization Across PDE Families [17.225653683970393]
We propose a unified framework for multi-operator few-shot learning, which aims to generalize to unseen PDE operators.<n>Our method integrates three key components: (i) multi-task self-supervised pretraining of a shared Fourier Neural Operator (FNO) encoder, (ii) text-conditioned operator embeddings derived from statistical summaries of input-output fields, and (iii) memory-augmented multimodal prompting.<n> Experiments on PDE benchmarks, including Darcy Flow and Navier Stokes variants, demonstrate that our model outperforms existing operator learning baselines in few-shot generalization.
arXiv Detail & Related papers (2025-08-02T06:00:01Z) - Principled Approaches for Extending Neural Architectures to Function Spaces for Operator Learning [78.88684753303794]
Deep learning has predominantly advanced through applications in computer vision and natural language processing.<n>Neural operators are a principled way to generalize neural networks to mappings between function spaces.<n>This paper identifies and distills the key principles for constructing practical implementations of mappings between infinite-dimensional function spaces.
arXiv Detail & Related papers (2025-06-12T17:59:31Z) - New universal operator approximation theorem for encoder-decoder architectures (Preprint) [0.0]
We present a novel universal operator approximation theorem for a broad class of encoder-decoder architectures.<n>In this study, we focus on approximating continuous operators in $mathcalC(mathcalX, mathcalY)$, where $mathcalX$ and $mathcalY$ are infinite-dimensional normed or metric spaces.
arXiv Detail & Related papers (2025-03-31T13:43:21Z) - Operator Learning of Lipschitz Operators: An Information-Theoretic Perspective [2.375038919274297]
This work addresses the complexity of neural operator approximations for the general class of Lipschitz continuous operators.
Our main contribution establishes lower bounds on the metric entropy of Lipschitz operators in two approximation settings.
It is shown that, regardless of the activation function used, neural operator architectures attaining an approximation accuracy $epsilon$ must have a size that is exponentially large in $epsilon-1$.
arXiv Detail & Related papers (2024-06-26T23:36:46Z) - MODNO: Multi Operator Learning With Distributed Neural Operators [0.8702432681310401]
The study of operator learning involves the utilization of neural networks to approximate operators.
Recent advances have led to the approximation of multiple operators using foundation models equipped with millions or billions of trainable parameters.
We present a novel distributed training approach aimed at enabling a single neural operator with significantly fewer parameters to tackle multi-operator learning challenges.
arXiv Detail & Related papers (2024-04-03T17:49:41Z) - Neural Operators with Localized Integral and Differential Kernels [77.76991758980003]
We present a principled approach to operator learning that can capture local features under two frameworks.
We prove that we obtain differential operators under an appropriate scaling of the kernel values of CNNs.
To obtain local integral operators, we utilize suitable basis representations for the kernels based on discrete-continuous convolutions.
arXiv Detail & Related papers (2024-02-26T18:59:31Z) - Multi-Operational Mathematical Derivations in Latent Space [16.255836734376206]
We introduce different multi-operational representation paradigms, modelling mathematical operations as explicit geometric transformations.
We construct a large-scale dataset comprising 1.7M derivation steps stemming from 61K premises and 6 operators.
We show that architectural choices can heavily affect the training dynamics, structural organisation, and generalisation of the latent space.
arXiv Detail & Related papers (2023-11-02T13:33:07Z) - Stabilizing Q-learning with Linear Architectures for Provably Efficient
Learning [53.17258888552998]
This work proposes an exploration variant of the basic $Q$-learning protocol with linear function approximation.
We show that the performance of the algorithm degrades very gracefully under a novel and more permissive notion of approximation error.
arXiv Detail & Related papers (2022-06-01T23:26:51Z) - Neural Operator: Learning Maps Between Function Spaces [75.93843876663128]
We propose a generalization of neural networks to learn operators, termed neural operators, that map between infinite dimensional function spaces.
We prove a universal approximation theorem for our proposed neural operator, showing that it can approximate any given nonlinear continuous operator.
An important application for neural operators is learning surrogate maps for the solution operators of partial differential equations.
arXiv Detail & Related papers (2021-08-19T03:56:49Z) - Statistically Meaningful Approximation: a Case Study on Approximating
Turing Machines with Transformers [50.85524803885483]
This work proposes a formal definition of statistically meaningful (SM) approximation which requires the approximating network to exhibit good statistical learnability.
We study SM approximation for two function classes: circuits and Turing machines.
arXiv Detail & Related papers (2021-07-28T04:28:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.