HyperLoRA for PDEs
- URL: http://arxiv.org/abs/2308.09290v1
- Date: Fri, 18 Aug 2023 04:29:48 GMT
- Title: HyperLoRA for PDEs
- Authors: Ritam Majumdar, Vishal Jadhav, Anirudh Deodhar, Shirish Karande,
Lovekesh Vig, Venkataramana Runkana
- Abstract summary: Physics-informed neural networks (PINNs) have been widely used to develop neural surrogates for solutions of Partial Differential Equations.
A drawback of PINNs is that they have to be retrained with every change in initial-boundary conditions and PDE coefficients.
The Hypernetwork, a model-based meta learning technique, takes in a parameterized task embedding as input and predicts the weights of PINN as output.
- Score: 7.898728380447954
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Physics-informed neural networks (PINNs) have been widely used to develop
neural surrogates for solutions of Partial Differential Equations. A drawback
of PINNs is that they have to be retrained with every change in
initial-boundary conditions and PDE coefficients. The Hypernetwork, a
model-based meta learning technique, takes in a parameterized task embedding as
input and predicts the weights of PINN as output. Predicting weights of a
neural network however, is a high-dimensional regression problem, and
hypernetworks perform sub-optimally while predicting parameters for large base
networks. To circumvent this issue, we use a low ranked adaptation (LoRA)
formulation to decompose every layer of the base network into low-ranked
tensors and use hypernetworks to predict the low-ranked tensors. Despite the
reduced dimensionality of the resulting weight-regression problem, LoRA-based
Hypernetworks violate the underlying physics of the given task. We demonstrate
that the generalization capabilities of LoRA-based hypernetworks drastically
improve when trained with an additional physics-informed loss component
(HyperPINN) to satisfy the governing differential equations. We observe that
LoRA-based HyperPINN training allows us to learn fast solutions for
parameterized PDEs like Burger's equation and Navier Stokes: Kovasznay flow,
while having an 8x reduction in prediction parameters on average without
compromising on accuracy when compared to all other baselines.
Related papers
- Gradient Descent Finds Over-Parameterized Neural Networks with Sharp Generalization for Nonparametric Regression: A Distribution-Free Analysis [19.988762532185884]
We show that, if the neural network is trained by GD with early stopping, then the trained network renders a sharp rate of the nonparametric regression risk of $cO(eps_n2)$.
It is remarked that our result does not require distributional assumptions on the training data.
arXiv Detail & Related papers (2024-11-05T08:43:54Z) - HyResPINNs: Adaptive Hybrid Residual Networks for Learning Optimal Combinations of Neural and RBF Components for Physics-Informed Modeling [22.689531776611084]
We present a new class of PINNs called HyResPINNs, which augment traditional PINNs with adaptive hybrid residual blocks.
A key feature of our method is the inclusion of adaptive combination parameters within each residual block.
We show that HyResPINNs are more robust to training point locations and neural network architectures than traditional PINNs.
arXiv Detail & Related papers (2024-10-04T16:21:14Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - NeuralStagger: Accelerating Physics-constrained Neural PDE Solver with
Spatial-temporal Decomposition [67.46012350241969]
This paper proposes a general acceleration methodology called NeuralStagger.
It decomposing the original learning tasks into several coarser-resolution subtasks.
We demonstrate the successful application of NeuralStagger on 2D and 3D fluid dynamics simulations.
arXiv Detail & Related papers (2023-02-20T19:36:52Z) - RBF-MGN:Solving spatiotemporal PDEs with Physics-informed Graph Neural
Network [4.425915683879297]
We propose a novel framework based on graph neural networks (GNNs) and radial basis function finite difference (RBF-FD)
RBF-FD is used to construct a high-precision difference format of the differential equations to guide model training.
We illustrate the generalizability, accuracy, and efficiency of the proposed algorithms on different PDE parameters.
arXiv Detail & Related papers (2022-12-06T10:08:02Z) - Neural Basis Functions for Accelerating Solutions to High Mach Euler
Equations [63.8376359764052]
We propose an approach to solving partial differential equations (PDEs) using a set of neural networks.
We regress a set of neural networks onto a reduced order Proper Orthogonal Decomposition (POD) basis.
These networks are then used in combination with a branch network that ingests the parameters of the prescribed PDE to compute a reduced order approximation to the PDE.
arXiv Detail & Related papers (2022-08-02T18:27:13Z) - Learning to Solve PDE-constrained Inverse Problems with Graph Networks [51.89325993156204]
In many application domains across science and engineering, we are interested in solving inverse problems with constraints defined by a partial differential equation (PDE)
Here we explore GNNs to solve such PDE-constrained inverse problems.
We demonstrate computational speedups of up to 90x using GNNs compared to principled solvers.
arXiv Detail & Related papers (2022-06-01T18:48:01Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - PhyCRNet: Physics-informed Convolutional-Recurrent Network for Solving
Spatiotemporal PDEs [8.220908558735884]
Partial differential equations (PDEs) play a fundamental role in modeling and simulating problems across a wide range of disciplines.
Recent advances in deep learning have shown the great potential of physics-informed neural networks (NNs) to solve PDEs as a basis for data-driven inverse analysis.
We propose the novel physics-informed convolutional-recurrent learning architectures (PhyCRNet and PhCRyNet-s) for solving PDEs without any labeled data.
arXiv Detail & Related papers (2021-06-26T22:22:19Z) - Quadratic Residual Networks: A New Class of Neural Networks for Solving
Forward and Inverse Problems in Physics Involving PDEs [2.8718027848324317]
quadratic residual networks (QRes) are a new type of parameter-efficient neural network architecture.
We show that QRes is especially powerful for solving forward and inverse physics problems involving partial differential equations (PDEs)
We empirically show that QRes shows faster convergence speed in terms of number of training epochs especially in learning complex patterns.
arXiv Detail & Related papers (2021-01-20T23:54:55Z) - Revisiting Initialization of Neural Networks [72.24615341588846]
We propose a rigorous estimation of the global curvature of weights across layers by approximating and controlling the norm of their Hessian matrix.
Our experiments on Word2Vec and the MNIST/CIFAR image classification tasks confirm that tracking the Hessian norm is a useful diagnostic tool.
arXiv Detail & Related papers (2020-04-20T18:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.