Data-Efficient Time-Dependent PDE Surrogates: Graph Neural Simulators vs. Neural Operators
- URL: http://arxiv.org/abs/2509.06154v2
- Date: Wed, 24 Sep 2025 18:01:13 GMT
- Title: Data-Efficient Time-Dependent PDE Surrogates: Graph Neural Simulators vs. Neural Operators
- Authors: Dibyajyoti Nayak, Somdatta Goswami,
- Abstract summary: We propose Neural Graph Simulators (GNS) as a principled surrogate modeling paradigm for time-dependent partial differential equations (PDEs)<n>GNS leverages message-passing combined with numerical time-stepping schemes to learn PDE dynamics by modeling the instantaneous time derivatives.<n>Results demonstrate that GNS is markedly more data-efficient, achieving less than 1% relative L2 error using only 3% of available trajectories.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Developing accurate, data-efficient surrogate models is central to advancing AI for Science. Neural operators (NOs), which approximate mappings between infinite-dimensional function spaces using conventional neural architectures, have gained popularity as surrogates for systems driven by partial differential equations (PDEs). However, their reliance on large datasets and limited ability to generalize in low-data regimes hinder their practical utility. We argue that these limitations arise from their global processing of data, which fails to exploit the local, discretized structure of physical systems. To address this, we propose Graph Neural Simulators (GNS) as a principled surrogate modeling paradigm for time-dependent PDEs. GNS leverages message-passing combined with numerical time-stepping schemes to learn PDE dynamics by modeling the instantaneous time derivatives. This design mimics traditional numerical solvers, enabling stable long-horizon rollouts and strong inductive biases that enhance generalization. We rigorously evaluate GNS on four canonical PDE systems: (1) 2D scalar Burgers', (2) 2D coupled Burgers', (3) 2D Allen-Cahn, and (4) 2D nonlinear shallow-water equations, comparing against state-of-the-art NOs including Deep Operator Network (DeepONet) and Fourier Neural Operator (FNO). Results demonstrate that GNS is markedly more data-efficient, achieving less than 1% relative L2 error using only 3% of available trajectories, and exhibits dramatically reduced error accumulation over time (82.5% lower autoregressive error than FNO, 99.9% lower than DeepONet). To choose the training data, we introduce a PCA combined with KMeans trajectory selection strategy. These findings provide compelling evidence that GNS, with its graph-based locality and solver-inspired design, is the most suitable and scalable surrogate modeling framework for AI-driven scientific discovery.
Related papers
- Stable spectral neural operator for learning stiff PDE systems from limited data [12.62991453201434]
We introduce an equation-free learning framework, namely, the Stable Spectral Neural Operator (SSNO)<n>SSNO embeds spectrally inspired structures in its architecture, yielding strong inductive biases for learning the underlying physics.<n>It achieves prediction errors one to two orders of magnitude lower than leading models.
arXiv Detail & Related papers (2025-12-12T16:09:38Z) - A data free neural operator enabling fast inference of 2D and 3D Navier Stokes equations [9.220574860416798]
We present a data-free operator network for the Navier Stokes equations that eliminates the need for paired solution data.<n>Across 2D benchmarks and 3D test cases, the method surpasses prior neural operators in accuracy and, for ensembles, achieves greater efficiency than conventional numerical solvers.
arXiv Detail & Related papers (2025-10-27T23:41:42Z) - Data-Augmented Few-Shot Neural Stencil Emulation for System Identification of Computer Models [13.794891110913776]
Partial equations (PDEs) underpin the modeling of many natural and engineered systems.<n>It can be convenient to express such models as neural PDEs by replacing part or all of the PDE's governing equations with a neural network representation.<n>We propose a more sample-efficient data-augmentation strategy for generating neural PDE training data from a computer model.
arXiv Detail & Related papers (2025-08-26T21:22:11Z) - ODE-GS: Latent ODEs for Dynamic Scene Extrapolation with 3D Gaussian Splatting [10.497667917243852]
ODE-GS is a novel method that unifies 3D Gaussian Splatting with latent neural ordinary differential equations (ODEs) to forecast dynamic 3D scenes.<n>Our results demonstrate that continuous-time latent dynamics are a powerful, practical route to prediction of complex 3D scenes.
arXiv Detail & Related papers (2025-06-05T18:02:30Z) - Enabling Automatic Differentiation with Mollified Graph Neural Operators [75.3183193262225]
We propose the mollified graph neural operator (mGNO), the first method to leverage automatic differentiation and compute emphexact gradients on arbitrary geometries.<n>For a PDE example on regular grids, mGNO paired with autograd reduced the L2 relative data error by 20x compared to finite differences.<n>It can also solve PDEs on unstructured point clouds seamlessly, using physics losses only, at resolutions vastly lower than those needed for finite differences to be accurate enough.
arXiv Detail & Related papers (2025-04-11T06:16:30Z) - Model-Agnostic Knowledge Guided Correction for Improved Neural Surrogate Rollout [3.006104092368596]
We propose a model-agnostic, RL based, cost-aware model which combines a neural surrogate, RL decision model, and a physics simulator to reduce surrogate rollout error significantly.<n>HyPER learns an intelligent policy that is adaptable to changing physical conditions and resistant to noise corruption.
arXiv Detail & Related papers (2025-03-13T05:00:23Z) - Solving Oscillator Ordinary Differential Equations in the Time Domain with High Performance via Soft-constrained Physics-informed Neural Network with Small Data [0.6446246430600296]
Physics-informed neural network (PINN) incorporates physical information and knowledge into network topology or computational processes as model priors.<n>This study aims to investigate the performance characteristics of the soft-constrained PINN method to solve typical linear and nonlinear ordinary differential equations.
arXiv Detail & Related papers (2024-08-19T13:02:06Z) - Learning time-dependent PDE via graph neural networks and deep operator
network for robust accuracy on irregular grids [14.93012615797081]
GraphDeepONet is an autoregressive model based on graph neural networks (GNNs)
It exhibits robust accuracy in predicting solutions compared to existing GNN-based PDE solver models.
Unlike traditional DeepONet and its variants, GraphDeepONet enables time extrapolation for time-dependent PDE solutions.
arXiv Detail & Related papers (2024-02-13T03:14:32Z) - Guaranteed Approximation Bounds for Mixed-Precision Neural Operators [83.64404557466528]
We build on intuition that neural operator learning inherently induces an approximation error.
We show that our approach reduces GPU memory usage by up to 50% and improves throughput by 58% with little or no reduction in accuracy.
arXiv Detail & Related papers (2023-07-27T17:42:06Z) - Training Deep Surrogate Models with Large Scale Online Learning [48.7576911714538]
Deep learning algorithms have emerged as a viable alternative for obtaining fast solutions for PDEs.
Models are usually trained on synthetic data generated by solvers, stored on disk and read back for training.
It proposes an open source online training framework for deep surrogate models.
arXiv Detail & Related papers (2023-06-28T12:02:27Z) - NUNO: A General Framework for Learning Parametric PDEs with Non-Uniform
Data [25.52271761404213]
We introduce the Non-Uniform Operator (NUNO) framework for efficient operator learning with non-uniform data.
We transform non-uniform data into uniform grids while effectively controlling error, thereby paralleling the speed and accuracy of non-uniform data.
Our framework has reduced error rates by up to 60% and enhanced training speeds by 2x to 30x.
arXiv Detail & Related papers (2023-05-30T02:34:10Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - Monte Carlo Neural PDE Solver for Learning PDEs via Probabilistic Representation [59.45669299295436]
We propose a Monte Carlo PDE solver for training unsupervised neural solvers.<n>We use the PDEs' probabilistic representation, which regards macroscopic phenomena as ensembles of random particles.<n>Our experiments on convection-diffusion, Allen-Cahn, and Navier-Stokes equations demonstrate significant improvements in accuracy and efficiency.
arXiv Detail & Related papers (2023-02-10T08:05:19Z) - Solving High-Dimensional PDEs with Latent Spectral Models [74.1011309005488]
We present Latent Spectral Models (LSM) toward an efficient and precise solver for high-dimensional PDEs.
Inspired by classical spectral methods in numerical analysis, we design a neural spectral block to solve PDEs in the latent space.
LSM achieves consistent state-of-the-art and yields a relative gain of 11.5% averaged on seven benchmarks.
arXiv Detail & Related papers (2023-01-30T04:58:40Z) - LordNet: An Efficient Neural Network for Learning to Solve Parametric Partial Differential Equations without Simulated Data [47.49194807524502]
We propose LordNet, a tunable and efficient neural network for modeling entanglements.
The experiments on solving Poisson's equation and (2D and 3D) Navier-Stokes equation demonstrate that the long-range entanglements can be well modeled by the LordNet.
arXiv Detail & Related papers (2022-06-19T14:41:08Z) - Truncated tensor Schatten p-norm based approach for spatiotemporal
traffic data imputation with complicated missing patterns [77.34726150561087]
We introduce four complicated missing patterns, including missing and three fiber-like missing cases according to the mode-drivenn fibers.
Despite nonity of the objective function in our model, we derive the optimal solutions by integrating alternating data-mputation method of multipliers.
arXiv Detail & Related papers (2022-05-19T08:37:56Z) - EIGNN: Efficient Infinite-Depth Graph Neural Networks [51.97361378423152]
Graph neural networks (GNNs) are widely used for modelling graph-structured data in numerous applications.
Motivated by this limitation, we propose a GNN model with infinite depth, which we call Efficient Infinite-Depth Graph Neural Networks (EIGNN)
We show that EIGNN has a better ability to capture long-range dependencies than recent baselines, and consistently achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T08:16:58Z) - Neural ODE Processes [64.10282200111983]
We introduce Neural ODE Processes (NDPs), a new class of processes determined by a distribution over Neural ODEs.
We show that our model can successfully capture the dynamics of low-dimensional systems from just a few data-points.
arXiv Detail & Related papers (2021-03-23T09:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.