Distributed computing for physics-based data-driven reduced modeling at scale: Application to a rotating detonation rocket engine
- URL: http://arxiv.org/abs/2407.09994v2
- Date: Sat, 19 Apr 2025 16:29:33 GMT
- Title: Distributed computing for physics-based data-driven reduced modeling at scale: Application to a rotating detonation rocket engine
- Authors: Ionut-Gabriel Farcas, Rayomand P. Gundevia, Ramakanth Munipalli, Karen E. Willcox,
- Abstract summary: This paper contributes a distributed memory algorithm that achieves fast and scalable construction of predictive physics-based ROMs trained from sparse datasets of extremely large state dimension.<n>We focus on a real-world three-dimensional RDRE for which one millisecond of simulated physical time requires one million core hours on a supercomputer.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: High-performance computing (HPC) has revolutionized our ability to perform detailed simulations of complex real-world processes. A prominent contemporary example is from aerospace propulsion, where HPC is used for rotating detonation rocket engine (RDRE) simulations in support of the design of next-generation rocket engines; however, these simulations take millions of core hours even on powerful supercomputers, which makes them impractical for engineering tasks like design exploration and risk assessment. Data-driven reduced-order models (ROMs) aim to address this limitation by constructing computationally cheap yet sufficiently accurate approximations that serve as surrogates for the high-fidelity model. This paper contributes a distributed memory algorithm that achieves fast and scalable construction of predictive physics-based ROMs trained from sparse datasets of extremely large state dimension. The algorithm learns structured physics-based ROMs that approximate the dynamical systems underlying those datasets.This enables model reduction for problems at a scale and complexity that exceeds the capabilities of standard, serial approaches. We demonstrate our algorithm's scalability using up to $2,048$ cores on the Frontera supercomputer at the Texas Advanced Computing Center. We focus on a real-world three-dimensional RDRE for which one millisecond of simulated physical time requires one million core hours on a supercomputer. Using a training dataset of $2,536$ snapshots each of state dimension $76$ million, our distributed algorithm enables the construction of a predictive data-driven reduced model in just $13$ seconds on $2,048$ cores on Frontera.
Related papers
- DrivAerStar: An Industrial-Grade CFD Dataset for Vehicle Aerodynamic Optimization [17.541117004160892]
Traditional approaches face an intractable trade-off: computationally expensive Computational Dynamics (CFD) simulations generated weeks per design iteration, or simplified models that sacrifice production-grade accuracy.<n>DrivAerStar is the first dataset bridging academic machine learning research and industrial CFD practice, establishing a new standard for data-driven aerodynamic optimization in automotive development.<n>Beyond automotive applications, DrivAerStar demonstrates a paradigm for integrating high-fidelity physics simulations with Artificial Intelligence (AI) across engineering disciplines where computational constraints currently limit innovation.
arXiv Detail & Related papers (2025-10-19T14:42:32Z) - Fourier Neural Operator based surrogates for $CO_2$ storage in realistic geologies [57.23978190717341]
We develop a Neural Operator (FNO) based model for real-time, high-resolution simulation of $CO$ plume migration.
The model is trained on a comprehensive dataset generated from realistic subsurface parameters.
We present various strategies for improving the reliability of predictions from the model, which is crucial while assessing actual geological sites.
arXiv Detail & Related papers (2025-03-14T02:58:24Z) - Open-Source High-Speed Flight Surrogate Modeling Framework [0.0]
High-speed flight vehicles, which travel much faster than the speed of sound, are crucial for national defense and space exploration.
accurately predicting their behavior under numerous, varied flight conditions is a challenge and often expensive.
The proposed approach involves creating smarter, more efficient machine learning models.
arXiv Detail & Related papers (2024-11-06T01:34:06Z) - Compilation of Trotter-Based Time Evolution for Partially Fault-Tolerant Quantum Computing Architecture [0.6449786007855248]
We present an efficient method for simulating the time evolution of the 2D Hubbard model Hamiltonian.
Our analysis reveals an acceleration of over 10 times compared to naive serial compilation.
For devices with a physical error rate of $p_rm phys = 10-4$, we estimate that approximately $6.5 times 104$ physical qubits are required to achieve faster ground state energy estimation.
arXiv Detail & Related papers (2024-08-27T10:07:34Z) - Physics-Informed Machine Learning Towards A Real-Time Spacecraft Thermal Simulator [15.313871831214902]
The PIML model or hybrid model presented here consists of a neural network which predicts reduced nodalizations given on-orbit thermal load conditions.
We compare the computational performance and accuracy of the hybrid model to a data-driven neural net model, and a high-fidelity finite-difference model of a prototype Earth-orbiting small spacecraft.
The PIML based active nodalization approach provides significantly better generalization than the neural net model and coarse mesh model, while reducing computing cost by up to 1.7x compared to the high-fidelity model.
arXiv Detail & Related papers (2024-07-08T16:38:52Z) - Scientific Machine Learning Based Reduced-Order Models for Plasma Turbulence Simulations [0.0]
This paper investigates non-intrusive Scientific Machine Learning (SciML) Reduced-Order Models (ROMs) for plasma turbulence simulations.
We focus on Operator Inference (OpInf) to build low-cost physics-based ROMs from data for such simulations.
arXiv Detail & Related papers (2024-01-11T15:20:06Z) - Waymax: An Accelerated, Data-Driven Simulator for Large-Scale Autonomous
Driving Research [76.93956925360638]
Waymax is a new data-driven simulator for autonomous driving in multi-agent scenes.
It runs entirely on hardware accelerators such as TPUs/GPUs and supports in-graph simulation for training.
We benchmark a suite of popular imitation and reinforcement learning algorithms with ablation studies on different design decisions.
arXiv Detail & Related papers (2023-10-12T20:49:15Z) - $\textit{FastSVD-ML-ROM}$: A Reduced-Order Modeling Framework based on
Machine Learning for Real-Time Applications [0.0]
High-fidelity numerical simulations constitute the backbone of engineering design.
Reduced order models (ROMs) are employed to approximate the high-fidelity solutions.
The present work proposes a new machine learning (ML) platform for the development of ROMs.
arXiv Detail & Related papers (2022-07-24T23:11:07Z) - Learning Large-scale Subsurface Simulations with a Hybrid Graph Network
Simulator [57.57321628587564]
We introduce Hybrid Graph Network Simulator (HGNS) for learning reservoir simulations of 3D subsurface fluid flows.
HGNS consists of a subsurface graph neural network (SGNN) to model the evolution of fluid flows, and a 3D-U-Net to model the evolution of pressure.
Using an industry-standard subsurface flow dataset (SPE-10) with 1.1 million cells, we demonstrate that HGNS is able to reduce the inference time up to 18 times compared to standard subsurface simulators.
arXiv Detail & Related papers (2022-06-15T17:29:57Z) - Deep Learning for Real Time Satellite Pose Estimation on Low Power Edge
TPU [58.720142291102135]
In this paper we propose a pose estimation software exploiting neural network architectures.
We show how low power machine learning accelerators could enable Artificial Intelligence exploitation in space.
arXiv Detail & Related papers (2022-04-07T08:53:18Z) - FastFlowNet: A Lightweight Network for Fast Optical Flow Estimation [81.76975488010213]
Dense optical flow estimation plays a key role in many robotic vision tasks.
Current networks often occupy large number of parameters and require heavy computation costs.
Our proposed FastFlowNet works in the well-known coarse-to-fine manner with following innovations.
arXiv Detail & Related papers (2021-03-08T03:09:37Z) - Data-Efficient Learning for Complex and Real-Time Physical Problem
Solving using Augmented Simulation [49.631034790080406]
We present a task for navigating a marble to the center of a circular maze.
We present a model that learns to move a marble in the complex environment within minutes of interacting with the real system.
arXiv Detail & Related papers (2020-11-14T02:03:08Z) - One-step regression and classification with crosspoint resistive memory
arrays [62.997667081978825]
High speed, low energy computing machines are in demand to enable real-time artificial intelligence at the edge.
One-step learning is supported by simulations of the prediction of the cost of a house in Boston and the training of a 2-layer neural network for MNIST digit recognition.
Results are all obtained in one computational step, thanks to the physical, parallel, and analog computing within the crosspoint array.
arXiv Detail & Related papers (2020-05-05T08:00:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.