Physics-informed Multiple-Input Operators for efficient dynamic response prediction of structures
- URL: http://arxiv.org/abs/2505.07090v1
- Date: Sun, 11 May 2025 18:45:58 GMT
- Title: Physics-informed Multiple-Input Operators for efficient dynamic response prediction of structures
- Authors: Bilal Ahmed, Yuqing Qiu, Diab W. Abueidda, Waleed El-Sekelly, Tarek Abdoun, Mostafa E. Mobasher,
- Abstract summary: MIONet predicts structural responses continuously over both space and time.<n>Model is validated on both a simple beam and the KW-51 bridge, achieving FEM level accuracy within seconds.
- Score: 0.07916635054977067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Finite element (FE) modeling is essential for structural analysis but remains computationally intensive, especially under dynamic loading. While operator learning models have shown promise in replicating static structural responses at FEM level accuracy, modeling dynamic behavior remains more challenging. This work presents a Multiple Input Operator Network (MIONet) that incorporates a second trunk network to explicitly encode temporal dynamics, enabling accurate prediction of structural responses under moving loads. Traditional DeepONet architectures using recurrent neural networks (RNNs) are limited by fixed time discretization and struggle to capture continuous dynamics. In contrast, MIONet predicts responses continuously over both space and time, removing the need for step wise modeling. It maps scalar inputs including load type, velocity, spatial mesh, and time steps to full field structural responses. To improve efficiency and enforce physical consistency, we introduce a physics informed loss based on dynamic equilibrium using precomputed mass, damping, and stiffness matrices, without solving the governing PDEs directly. Further, a Schur complement formulation reduces the training domain, significantly cutting computational costs while preserving global accuracy. The model is validated on both a simple beam and the KW-51 bridge, achieving FEM level accuracy within seconds. Compared to GRU based DeepONet, our model offers comparable accuracy with improved temporal continuity and over 100 times faster inference, making it well suited for real-time structural monitoring and digital twin applications.
Related papers
- Iterative Pretraining Framework for Interatomic Potentials [46.53683458224917]
We propose Iterative Pretraining for Interatomic Potentials (IPIP) to improve predictive performance of MLIP models.<n>IPIP incorporates a forgetting mechanism to prevent iterative training from converging to suboptimal local minima.<n>Compared to general-purpose force fields, this approach achieves over 80% reduction in prediction error and up to 4x speedup in the challenging Mo-S-O system.
arXiv Detail & Related papers (2025-07-27T03:59:41Z) - OmniFluids: Unified Physics Pre-trained Modeling of Fluid Dynamics [25.066485418709114]
We introduce OmniFluids, a unified physics pre-trained operator learning framework.<n>It integrates physics-only pre-training, coarse-grid operator distillation, and few-shot fine-tuning.<n>It significantly outperforms state-of-the-art AI-driven methods in flow field reconstruction and turbulence statistics accuracy.
arXiv Detail & Related papers (2025-06-12T16:23:02Z) - Hybrid machine learning models based on physical patterns to accelerate CFD simulations: a short guide on autoregressive models [3.780691701083858]
This study presents an innovative integration of High-Order Singular Value Decomposition with Long Short-Term Memory (LSTM) architectures to address the complexities of reduced-order modeling (ROM) in fluid dynamics.<n>The methodology is tested across numerical and experimental data sets, including two- and three-dimensional (2D and 3D) cylinder wake flows, spanning both laminar and turbulent regimes.<n>The results demonstrate that HOSVD outperforms SVD in all tested scenarios, as evidenced by using different error metrics.
arXiv Detail & Related papers (2025-04-09T10:56:03Z) - Physics-Informed Latent Neural Operator for Real-time Predictions of Complex Physical Systems [0.0]
We propose PI-Latent-NO, a physics-informed latent neural operator framework that integrates governing physics directly into the learning process.<n>Our architecture features two coupled DeepONets trained end-to-end: a Latent-DeepONet that learns a low-dimensional representation of the solution, and a Reconstruction-DeepONet that maps this latent representation back to the physical space.
arXiv Detail & Related papers (2025-01-14T20:38:30Z) - Physics-Informed Regularization for Domain-Agnostic Dynamical System Modeling [41.82469276824927]
We present a framework that achieves high-precision modeling for a wide range of dynamical systems.
It helps preserve energies for conservative systems while serving as a strong inductive bias for non-conservative, reversible systems.
By integrating the TRS loss within neural ordinary differential equation models, the proposed model TREAT demonstrates superior performance on diverse physical systems.
arXiv Detail & Related papers (2024-10-08T21:04:01Z) - Temporally-Consistent Koopman Autoencoders for Forecasting Dynamical Systems [38.36312939874359]
We introduce the Temporally-Consistent Koopman Autoencoder (tcKAE)<n>tcKAE generates accurate long-term predictions even with limited and noisy training data.<n>We demonstrate tcKAE's superior performance over state-of-the-art KAE models across a variety of test cases.
arXiv Detail & Related papers (2024-03-19T00:48:25Z) - Physics-Informed Machine Learning for Seismic Response Prediction OF Nonlinear Steel Moment Resisting Frame Structures [6.483318568088176]
PiML method integrates scientific principles and physical laws into deep neural networks to model seismic responses of nonlinear structures.
Manipulating the equation of motion helps learn system nonlinearities and confines solutions within physically interpretable results.
Result handles complex data better than existing physics-guided LSTM models and outperforms other non-physics data-driven networks.
arXiv Detail & Related papers (2024-02-28T02:16:03Z) - Analyzing and Improving the Training Dynamics of Diffusion Models [36.37845647984578]
We identify and rectify several causes for uneven and ineffective training in the popular ADM diffusion model architecture.
We find that systematic application of this philosophy eliminates the observed drifts and imbalances, resulting in considerably better networks at equal computational complexity.
arXiv Detail & Related papers (2023-12-05T11:55:47Z) - Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at Scale [54.15522908057831]
We propose an adapted version of the computationally-Mixer for STTD forecast at scale.
Our results surprisingly show that this simple-yeteffective solution can rival SOTA baselines when tested on several traffic benchmarks.
Our findings contribute to the exploration of simple-yet-effective models for real-world STTD forecasting.
arXiv Detail & Related papers (2023-07-04T05:19:19Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - Real-time Neural-MPC: Deep Learning Model Predictive Control for
Quadrotors and Agile Robotic Platforms [59.03426963238452]
We present Real-time Neural MPC, a framework to efficiently integrate large, complex neural network architectures as dynamics models within a model-predictive control pipeline.
We show the feasibility of our framework on real-world problems by reducing the positional tracking error by up to 82% when compared to state-of-the-art MPC approaches without neural network dynamics.
arXiv Detail & Related papers (2022-03-15T09:38:15Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - AntiDote: Attention-based Dynamic Optimization for Neural Network
Runtime Efficiency [42.00372941618975]
We propose a dynamic CNN optimization framework based on the neural network attention mechanism.
Our method could bring 37.4% to 54.5% FLOPs reduction with negligible accuracy drop on various test networks.
arXiv Detail & Related papers (2020-08-14T18:48:13Z) - Liquid Time-constant Networks [117.57116214802504]
We introduce a new class of time-continuous recurrent neural network models.
Instead of declaring a learning system's dynamics by implicit nonlinearities, we construct networks of linear first-order dynamical systems.
These neural networks exhibit stable and bounded behavior, yield superior expressivity within the family of neural ordinary differential equations.
arXiv Detail & Related papers (2020-06-08T09:53:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.