Revisiting Orbital Minimization Method for Neural Operator Decomposition
- URL: http://arxiv.org/abs/2510.21952v1
- Date: Fri, 24 Oct 2025 18:26:18 GMT
- Title: Revisiting Orbital Minimization Method for Neural Operator Decomposition
- Authors: J. Jon Ryu, Samuel Zhou, Gregory W. Wornell,
- Abstract summary: We revisit a classical optimization framework known as the emphorbital method (OMM) originally proposed in the 1990s for solving eigenvalue problems in computational chemistry.<n>We adapt this framework to train neural networks to decompose positive semidefinite operators, and demonstrate its practical advantages across a range of benchmark tasks.
- Score: 19.86950069790711
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Spectral decomposition of linear operators plays a central role in many areas of machine learning and scientific computing. Recent work has explored training neural networks to approximate eigenfunctions of such operators, enabling scalable approaches to representation learning, dynamical systems, and partial differential equations (PDEs). In this paper, we revisit a classical optimization framework from the computational physics literature known as the \emph{orbital minimization method} (OMM), originally proposed in the 1990s for solving eigenvalue problems in computational chemistry. We provide a simple linear-algebraic proof of the consistency of the OMM objective, and reveal connections between this method and several ideas that have appeared independently across different domains. Our primary goal is to justify its broader applicability in modern learning pipelines. We adapt this framework to train neural networks to decompose positive semidefinite operators, and demonstrate its practical advantages across a range of benchmark tasks. Our results highlight how revisiting classical numerical methods through the lens of modern theory and computation can provide not only a principled approach for deploying neural networks in numerical simulation, but also effective and scalable tools for machine learning.
Related papers
- Tackling multiphysics problems via finite element-guided physics-informed operator learning [0.0]
This work presents a finite element-guided physics-informed operator learning framework for multiphysics problems.<n>The proposed framework learns a mapping from the input parameter space to the solution space with a weighted residual formulation based on the finite element method.<n>The present framework for multiphysics problems is verified on nonlinear thermo-mechanical problems.
arXiv Detail & Related papers (2026-03-02T03:52:51Z) - Soliton profiles: Classical Numerical Schemes vs. Neural Network - Based Solvers [0.24999074238880484]
We present a comparative study of classical numerical solvers and neural network-based methods.<n>We confirm that classical approaches retain high-order accuracy and strong computational efficiency for single-instance problems.<n>For single-instance computations, however, the accuracy of operator-learning methods remains lower than that of classical methods or PINNs.
arXiv Detail & Related papers (2025-12-31T05:13:16Z) - An Evolutionary Multi-objective Optimization for Replica-Exchange-based Physics-informed Operator Learning Network [7.1950116347185995]
We propose an evolutionary Multi-objective Optimization for Replica-based Physics-informed Operator learning Network.<n>Our framework consistently outperforms the general operator learning methods in accuracy, noise, and the ability to quantify uncertainty.
arXiv Detail & Related papers (2025-08-31T02:17:59Z) - Efficient Parametric SVD of Koopman Operator for Stochastic Dynamical Systems [51.54065545849027]
The Koopman operator provides a principled framework for analyzing nonlinear dynamical systems.<n>VAMPnet and DPNet have been proposed to learn the leading singular subspaces of the Koopman operator.<n>We propose a scalable and conceptually simple method for learning the top-$k$ singular functions of the Koopman operator.
arXiv Detail & Related papers (2025-07-09T18:55:48Z) - Principled Approaches for Extending Neural Architectures to Function Spaces for Operator Learning [78.88684753303794]
Deep learning has predominantly advanced through applications in computer vision and natural language processing.<n>Neural operators are a principled way to generalize neural networks to mappings between function spaces.<n>This paper identifies and distills the key principles for constructing practical implementations of mappings between infinite-dimensional function spaces.
arXiv Detail & Related papers (2025-06-12T17:59:31Z) - From Theory to Application: A Practical Introduction to Neural Operators in Scientific Computing [0.0]
The study covers foundational models such as Deep Operator Networks (DeepONet) and Principal Component Analysis-based Neural Networks (PCANet)<n>The review delves into applying neural operators as surrogates in Bayesian inference problems, showcasing their effectiveness in accelerating posterior inference while maintaining accuracy.<n>It outlines emerging strategies to address these issues, such as residual-based error correction and multi-level training.
arXiv Detail & Related papers (2025-03-07T17:25:25Z) - Mechanistic Neural Networks for Scientific Machine Learning [58.99592521721158]
We present Mechanistic Neural Networks, a neural network design for machine learning applications in the sciences.
It incorporates a new Mechanistic Block in standard architectures to explicitly learn governing differential equations as representations.
Central to our approach is a novel Relaxed Linear Programming solver (NeuRLP) inspired by a technique that reduces solving linear ODEs to solving linear programs.
arXiv Detail & Related papers (2024-02-20T15:23:24Z) - Approximating Numerical Fluxes Using Fourier Neural Operators for Hyperbolic Conservation Laws [7.438389089520601]
neural network-based methods, such as physics-informed neural networks (PINNs) and neural operators, exhibit deficiencies in robustness and generalization.
In this study, we focus on hyperbolic conservation laws by replacing traditional numerical flux with neural operators.
Our approach combines the strengths of both traditional numerical schemes and FNOs, outperforming standard FNO methods in several respects.
arXiv Detail & Related papers (2024-01-03T15:16:25Z) - Efficient and Flexible Neural Network Training through Layer-wise Feedback Propagation [49.44309457870649]
Layer-wise Feedback feedback (LFP) is a novel training principle for neural network-like predictors.<n>LFP decomposes a reward to individual neurons based on their respective contributions.<n>Our method then implements a greedy reinforcing approach helpful parts of the network and weakening harmful ones.
arXiv Detail & Related papers (2023-08-23T10:48:28Z) - Deep Efficient Continuous Manifold Learning for Time Series Modeling [11.876985348588477]
A symmetric positive definite matrix is being studied in computer vision, signal processing, and medical image analysis.
In this paper, we propose a framework to exploit a diffeomorphism mapping between Riemannian manifold and a Cholesky space.
For dynamic modeling of time-series data, we devise a continuous manifold learning method by systematically integrating a manifold ordinary differential equation and a gated recurrent neural network.
arXiv Detail & Related papers (2021-12-03T01:38:38Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.