Reduced-Order Neural Operators: Learning Lagrangian Dynamics on Highly Sparse Graphs
- URL: http://arxiv.org/abs/2407.03925v1
- Date: Thu, 4 Jul 2024 13:37:26 GMT
- Title: Reduced-Order Neural Operators: Learning Lagrangian Dynamics on Highly Sparse Graphs
- Authors: Hrishikesh Viswanath, Yue Chang, Julius Berner, Peter Yichen Chen, Aniket Bera,
- Abstract summary: We present a neural operator architecture to simulate Lagrangian dynamics, such as fluid flow, granular flows, and elastoplasticity.
Our model, GIOROM, learns temporal dynamics within a reduced-order setting, capturing spatial features from a highly sparse graph representation.
We show that point clouds of the order of 100,000 points can be inferred from sparse graphs with $sim$1000 points, with negligible change in time.
- Score: 20.271792055491662
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a neural operator architecture to simulate Lagrangian dynamics, such as fluid flow, granular flows, and elastoplasticity. Traditional numerical methods, such as the finite element method (FEM), suffer from long run times and large memory consumption. On the other hand, approaches based on graph neural networks are faster but still suffer from long computation times on dense graphs, which are often required for high-fidelity simulations. Our model, GIOROM or Graph Interaction Operator for Reduced-Order Modeling, learns temporal dynamics within a reduced-order setting, capturing spatial features from a highly sparse graph representation of the input and generalizing to arbitrary spatial locations during inference. The model is geometry-aware and discretization-agnostic and can generalize to different initial conditions, velocities, and geometries after training. We show that point clouds of the order of 100,000 points can be inferred from sparse graphs with $\sim$1000 points, with negligible change in computation time. We empirically evaluate our model on elastic solids, Newtonian fluids, Non-Newtonian fluids, Drucker-Prager granular flows, and von Mises elastoplasticity. On these benchmarks, our approach results in a 25$\times$ speedup compared to other neural network-based physics simulators while delivering high-fidelity predictions of complex physical systems and showing better performance on most benchmarks. The code and the demos are provided at https://github.com/HrishikeshVish/GIOROM.
Related papers
- Three-dimensional granular flow simulation using graph neural
network-based learned simulator [2.153852088624324]
We use a graph neural network (GNN) to develop a simulator for granular flows.
The simulator reproduces the overall behaviors of column collapses with various aspect ratios.
The speed of GNS outperforms high-fidelity numerical simulators by 300 times.
arXiv Detail & Related papers (2023-11-13T15:54:09Z) - Geometry-Informed Neural Operator for Large-Scale 3D PDEs [76.06115572844882]
We propose the geometry-informed neural operator (GINO) to learn the solution operator of large-scale partial differential equations.
We successfully trained GINO to predict the pressure on car surfaces using only five hundred data points.
arXiv Detail & Related papers (2023-09-01T16:59:21Z) - Graph Neural Network-based surrogate model for granular flows [2.153852088624324]
Granular flow dynamics is crucial for assessing various geotechnical risks, including landslides and debris flows.
Traditional continuum and discrete numerical methods are limited by their computational cost in simulating large-scale systems.
We develop a graph neural network-based simulator (GNS) that takes the current state of granular flow and predicts the next state using explicit integration by learning the local interaction laws.
arXiv Detail & Related papers (2023-05-09T07:28:12Z) - Learning Large-scale Subsurface Simulations with a Hybrid Graph Network
Simulator [57.57321628587564]
We introduce Hybrid Graph Network Simulator (HGNS) for learning reservoir simulations of 3D subsurface fluid flows.
HGNS consists of a subsurface graph neural network (SGNN) to model the evolution of fluid flows, and a 3D-U-Net to model the evolution of pressure.
Using an industry-standard subsurface flow dataset (SPE-10) with 1.1 million cells, we demonstrate that HGNS is able to reduce the inference time up to 18 times compared to standard subsurface simulators.
arXiv Detail & Related papers (2022-06-15T17:29:57Z) - Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs [65.18780403244178]
We propose a continuous model to forecast Multivariate Time series with dynamic Graph neural Ordinary Differential Equations (MTGODE)
Specifically, we first abstract multivariate time series into dynamic graphs with time-evolving node features and unknown graph structures.
Then, we design and solve a neural ODE to complement missing graph topologies and unify both spatial and temporal message passing.
arXiv Detail & Related papers (2022-02-17T02:17:31Z) - Deep Surrogate for Direct Time Fluid Dynamics [44.62475518267084]
Graph Neural Networks (GNN) can address the specificity of the irregular meshes commonly used in CFD simulations.
We present our ongoing work to design a novel direct time GNN architecture for irregular meshes.
arXiv Detail & Related papers (2021-12-16T10:08:20Z) - Simulating Continuum Mechanics with Multi-Scale Graph Neural Networks [0.17205106391379021]
We introduce MultiScaleGNN, a network multi-scale neural graph model for learning to infer unsteady mechanics.
We show that the proposed model can generalise from uniform advection fields to high-gradient fields on complex domains at test time and infer long-term Navier-Stokes solutions within a range of Reynolds numbers.
arXiv Detail & Related papers (2021-06-09T08:37:38Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Liquid Time-constant Networks [117.57116214802504]
We introduce a new class of time-continuous recurrent neural network models.
Instead of declaring a learning system's dynamics by implicit nonlinearities, we construct networks of linear first-order dynamical systems.
These neural networks exhibit stable and bounded behavior, yield superior expressivity within the family of neural ordinary differential equations.
arXiv Detail & Related papers (2020-06-08T09:53:35Z) - Learning to Simulate Complex Physics with Graph Networks [68.43901833812448]
We present a machine learning framework and model implementation that can learn to simulate a wide variety of challenging physical domains.
Our framework---which we term "Graph Network-based Simulators" (GNS)--represents the state of a physical system with particles, expressed as nodes in a graph, and computes dynamics via learned message-passing.
Our results show that our model can generalize from single-timestep predictions with thousands of particles during training, to different initial conditions, thousands of timesteps, and at least an order of magnitude more particles at test time.
arXiv Detail & Related papers (2020-02-21T16:44:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.