Generalized Lagrangian Neural Networks
- URL: http://arxiv.org/abs/2401.03728v2
- Date: Tue, 9 Jan 2024 11:24:16 GMT
- Title: Generalized Lagrangian Neural Networks
- Authors: Shanshan Xiao, Jiawei Zhang, Yifa Tang
- Abstract summary: We introduce a groundbreaking extension (Genralized Lagrangian Neural Networks) to Lagrangian Neural Networks (LNNs)
By leveraging the foundational importance of the Lagrangian within Lagrange's equations, we formulate the model based on the generalized Lagrange's equation.
This modification not only enhances prediction accuracy but also guarantees Lagrangian representation in non-conservative systems.
- Score: 8.065464912030352
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Incorporating neural networks for the solution of Ordinary Differential
Equations (ODEs) represents a pivotal research direction within computational
mathematics. Within neural network architectures, the integration of the
intrinsic structure of ODEs offers advantages such as enhanced predictive
capabilities and reduced data utilization. Among these structural ODE forms,
the Lagrangian representation stands out due to its significant physical
underpinnings. Building upon this framework, Bhattoo introduced the concept of
Lagrangian Neural Networks (LNNs). Then in this article, we introduce a
groundbreaking extension (Genralized Lagrangian Neural Networks) to Lagrangian
Neural Networks (LNNs), innovatively tailoring them for non-conservative
systems. By leveraging the foundational importance of the Lagrangian within
Lagrange's equations, we formulate the model based on the generalized
Lagrange's equation. This modification not only enhances prediction accuracy
but also guarantees Lagrangian representation in non-conservative systems.
Furthermore, we perform various experiments, encompassing 1-dimensional and
2-dimensional examples, along with an examination of the impact of network
parameters, which proved the superiority of Generalized Lagrangian Neural
Networks(GLNNs).
Related papers
- Theoretical characterisation of the Gauss-Newton conditioning in Neural Networks [5.851101657703105]
We take a first step towards theoretically characterizing the conditioning of the Gauss-Newton (GN) matrix in neural networks.
We establish tight bounds on the condition number of the GN in deep linear networks of arbitrary depth and width.
We expand the analysis to further architectural components, such as residual connections and convolutional layers.
arXiv Detail & Related papers (2024-11-04T14:56:48Z) - Novel Kernel Models and Exact Representor Theory for Neural Networks Beyond the Over-Parameterized Regime [52.00917519626559]
This paper presents two models of neural-networks and their training applicable to neural networks of arbitrary width, depth and topology.
We also present an exact novel representor theory for layer-wise neural network training with unregularized gradient descent in terms of a local-extrinsic neural kernel (LeNK)
This representor theory gives insight into the role of higher-order statistics in neural network training and the effect of kernel evolution in neural-network kernel models.
arXiv Detail & Related papers (2024-05-24T06:30:36Z) - Mechanistic Neural Networks for Scientific Machine Learning [58.99592521721158]
We present Mechanistic Neural Networks, a neural network design for machine learning applications in the sciences.
It incorporates a new Mechanistic Block in standard architectures to explicitly learn governing differential equations as representations.
Central to our approach is a novel Relaxed Linear Programming solver (NeuRLP) inspired by a technique that reduces solving linear ODEs to solving linear programs.
arXiv Detail & Related papers (2024-02-20T15:23:24Z) - Fully Differentiable Lagrangian Convolutional Neural Network for
Continuity-Consistent Physics-Informed Precipitation Nowcasting [0.0]
We present a convolutional neural network model for precipitation nowcasting that combines data-driven learning with physics-informed domain knowledge.
We propose LUPIN, a Lagrangian Double U-Net for Physics-Informed Nowcasting, that draws from existing extrapolation-based nowcasting methods.
Based on our evaluation, LUPIN matches and exceeds the performance of the chosen benchmark, opening the door for other Lagrangian machine learning models.
arXiv Detail & Related papers (2024-02-16T15:13:30Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - Lagrangian Density Space-Time Deep Neural Network Topology [0.0]
We have proposed a "Lagrangian Density Space-Time Deep Neural Networks" (LDDNN) topology.
It is qualified for unsupervised training and learning to predict the dynamics of underlying physical science governed phenomena.
This article will discuss statistical physics interpretation of neural networks in the Lagrangian and Hamiltonian domains.
arXiv Detail & Related papers (2022-06-30T03:29:35Z) - Universal approximation property of invertible neural networks [76.95927093274392]
Invertible neural networks (INNs) are neural network architectures with invertibility by design.
Thanks to their invertibility and the tractability of Jacobian, INNs have various machine learning applications such as probabilistic modeling, generative modeling, and representation learning.
arXiv Detail & Related papers (2022-04-15T10:45:26Z) - Stability of Algebraic Neural Networks to Small Perturbations [179.55535781816343]
Algebraic neural networks (AlgNNs) are composed of a cascade of layers each one associated to and algebraic signal model.
We show how any architecture that uses a formal notion of convolution can be stable beyond particular choices of the shift operator.
arXiv Detail & Related papers (2020-10-22T09:10:16Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - A Lagrangian Approach to Information Propagation in Graph Neural
Networks [21.077268852378385]
In this paper, we propose a novel approach to the state computation and the learning algorithm for Graph Neural Network (GNN) models.
The state convergence procedure is implicitly expressed by the constraint satisfaction mechanism and does not require a separate iterative phase for each epoch of the learning procedure.
In fact, the computational structure is based on the search for saddle points of the Lagrangian in the adjoint space composed of weights, neural outputs (node states) and Lagrange multipliers.
arXiv Detail & Related papers (2020-02-18T16:13:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.