A Sequential Meta-Transfer (SMT) Learning to Combat Complexities of
Physics-Informed Neural Networks: Application to Composites Autoclave
Processing
- URL: http://arxiv.org/abs/2308.06447v1
- Date: Sat, 12 Aug 2023 02:46:54 GMT
- Title: A Sequential Meta-Transfer (SMT) Learning to Combat Complexities of
Physics-Informed Neural Networks: Application to Composites Autoclave
Processing
- Authors: Milad Ramezankhani and Abbas S. Milani
- Abstract summary: PINNs have gained popularity in solving nonlinear partial differential equations.
PINNs are designed to approximate a specific realization of a given PDE system.
They lack the necessary generalizability to efficiently adapt to new system configurations.
- Score: 1.6317061277457001
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Physics-Informed Neural Networks (PINNs) have gained popularity in solving
nonlinear partial differential equations (PDEs) via integrating physical laws
into the training of neural networks, making them superior in many scientific
and engineering applications. However, conventional PINNs still fall short in
accurately approximating the solution of complex systems with strong
nonlinearity, especially in long temporal domains. Besides, since PINNs are
designed to approximate a specific realization of a given PDE system, they lack
the necessary generalizability to efficiently adapt to new system
configurations. This entails computationally expensive re-training from scratch
for any new change in the system. To address these shortfalls, in this work a
novel sequential meta-transfer (SMT) learning framework is proposed, offering a
unified solution for both fast training and efficient adaptation of PINNs in
highly nonlinear systems with long temporal domains. Specifically, the
framework decomposes PDE's time domain into smaller time segments to create
"easier" PDE problems for PINNs training. Then for each time interval, a
meta-learner is assigned and trained to achieve an optimal initial state for
rapid adaptation to a range of related tasks. Transfer learning principles are
then leveraged across time intervals to further reduce the computational
cost.Through a composites autoclave processing case study, it is shown that SMT
is clearly able to enhance the adaptability of PINNs while significantly
reducing computational cost, by a factor of 100.
Related papers
- Efficient Discrete Physics-informed Neural Networks for Addressing
Evolutionary Partial Differential Equations [7.235476098729406]
Physics-informed neural networks (PINNs) have shown promising potential for solving partial differential equations (PDEs) using deep learning.
PINNs may violate the temporal causality property since all the temporal features in the PINNs loss are trained simultaneously.
This paper proposes to use implicit time differencing schemes to enforce temporal causality, and use transfer learning to sequentially update the PINNs in space as surrogates for PDE solutions in different time frames.
arXiv Detail & Related papers (2023-12-22T11:09:01Z) - PINNsFormer: A Transformer-Based Framework For Physics-Informed Neural Networks [22.39904196850583]
Physics-Informed Neural Networks (PINNs) have emerged as a promising deep learning framework for approximating numerical solutions to partial differential equations (PDEs)
We introduce a novel Transformer-based framework, termed PINNsFormer, designed to address this limitation.
PINNsFormer achieves superior generalization ability and accuracy across various scenarios, including PINNs failure modes and high-dimensional PDEs.
arXiv Detail & Related papers (2023-07-21T18:06:27Z) - A unified scalable framework for causal sweeping strategies for
Physics-Informed Neural Networks (PINNs) and their temporal decompositions [22.514769448363754]
Training challenges in PINNs and XPINNs for time-dependent PDEs are discussed.
We propose a new stacked-decomposition method that bridges the gap between PINNs and XPINNs.
We also formulate a new time-sweeping collocation point algorithm inspired by the previous PINNs causality.
arXiv Detail & Related papers (2023-02-28T01:19:21Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Separable PINN: Mitigating the Curse of Dimensionality in
Physics-Informed Neural Networks [6.439575695132489]
Physics-informed neural networks (PINNs) have emerged as new data-driven PDE solvers for both forward and inverse problems.
We demonstrate that the computations in automatic differentiation (AD) can be significantly reduced by leveraging forward-mode AD when training PINN.
We propose a network architecture, called separable PINN (SPINN), which can facilitate forward-mode AD for more efficient computation.
arXiv Detail & Related papers (2022-11-16T08:46:52Z) - Auto-PINN: Understanding and Optimizing Physics-Informed Neural
Architecture [77.59766598165551]
Physics-informed neural networks (PINNs) are revolutionizing science and engineering practice by bringing together the power of deep learning to bear on scientific computation.
Here, we propose Auto-PINN, which employs Neural Architecture Search (NAS) techniques to PINN design.
A comprehensive set of pre-experiments using standard PDE benchmarks allows us to probe the structure-performance relationship in PINNs.
arXiv Detail & Related papers (2022-05-27T03:24:31Z) - Revisiting PINNs: Generative Adversarial Physics-informed Neural
Networks and Point-weighting Method [70.19159220248805]
Physics-informed neural networks (PINNs) provide a deep learning framework for numerically solving partial differential equations (PDEs)
We propose the generative adversarial neural network (GA-PINN), which integrates the generative adversarial (GA) mechanism with the structure of PINNs.
Inspired from the weighting strategy of the Adaboost method, we then introduce a point-weighting (PW) method to improve the training efficiency of PINNs.
arXiv Detail & Related papers (2022-05-18T06:50:44Z) - Improved Training of Physics-Informed Neural Networks with Model
Ensembles [81.38804205212425]
We propose to expand the solution interval gradually to make the PINN converge to the correct solution.
All ensemble members converge to the same solution in the vicinity of observed data.
We show experimentally that the proposed method can improve the accuracy of the found solution.
arXiv Detail & Related papers (2022-04-11T14:05:34Z) - Characterizing possible failure modes in physics-informed neural
networks [55.83255669840384]
Recent work in scientific machine learning has developed so-called physics-informed neural network (PINN) models.
We demonstrate that, while existing PINN methodologies can learn good models for relatively trivial problems, they can easily fail to learn relevant physical phenomena even for simple PDEs.
We show that these possible failure modes are not due to the lack of expressivity in the NN architecture, but that the PINN's setup makes the loss landscape very hard to optimize.
arXiv Detail & Related papers (2021-09-02T16:06:45Z) - Physics-informed attention-based neural network for solving non-linear
partial differential equations [6.103365780339364]
Physics-Informed Neural Networks (PINNs) have enabled significant improvements in modelling physical processes.
PINNs are based on simple architectures, and learn the behavior of complex physical systems by optimizing the network parameters to minimize the residual of the underlying PDE.
Here, we address the question of which network architectures are best suited to learn the complex behavior of non-linear PDEs.
arXiv Detail & Related papers (2021-05-17T14:29:08Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.