Kohn-Sham equations as regularizer: building prior knowledge into
machine-learned physics
- URL: http://arxiv.org/abs/2009.08551v2
- Date: Tue, 17 Nov 2020 20:03:52 GMT
- Title: Kohn-Sham equations as regularizer: building prior knowledge into
machine-learned physics
- Authors: Li Li, Stephan Hoyer, Ryan Pederson, Ruoxi Sun, Ekin D. Cubuk, Patrick
Riley, Kieron Burke
- Abstract summary: We show that solving the Kohn-Sham equations when training neural networks for the exchange-correlation functional provides an implicit regularization that greatly improves generalization.
Our models also generalize to unseen types of molecules and overcome self-interaction error.
- Score: 13.572347341147282
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Including prior knowledge is important for effective machine learning models
in physics, and is usually achieved by explicitly adding loss terms or
constraints on model architectures. Prior knowledge embedded in the physics
computation itself rarely draws attention. We show that solving the Kohn-Sham
equations when training neural networks for the exchange-correlation functional
provides an implicit regularization that greatly improves generalization. Two
separations suffice for learning the entire one-dimensional H$_2$ dissociation
curve within chemical accuracy, including the strongly correlated region. Our
models also generalize to unseen types of molecules and overcome
self-interaction error.
Related papers
- Physics-informed machine learning of the correlation functions in bulk
fluids [2.1255150235172837]
The Ornstein-Zernike (OZ) equation is the fundamental equation for pair correlation function computations in the modern integral equation theory for liquids.
In this work, machine learning models, notably physics-informed neural networks and physics-informed neural operator networks, are explored to solve the OZ equation.
arXiv Detail & Related papers (2023-09-02T00:11:48Z) - Neural Astrophysical Wind Models [0.0]
We show that deep neural networks embedded as individual terms in the governing coupled ordinary differential equations (ODEs) can robustly discover both of these physics.
We optimize a loss function based on the Mach number, rather than the explicitly solved-for 3 conserved variables, and apply a penalty term towards near-diverging solutions.
This work further highlights the feasibility of neural ODEs as a promising discovery tool with mechanistic interpretability for non-linear inverse problems.
arXiv Detail & Related papers (2023-06-20T16:37:57Z) - Deep learning applied to computational mechanics: A comprehensive
review, state of the art, and the classics [77.34726150561087]
Recent developments in artificial neural networks, particularly deep learning (DL), are reviewed in detail.
Both hybrid and pure machine learning (ML) methods are discussed.
History and limitations of AI are recounted and discussed, with particular attention at pointing out misstatements or misconceptions of the classics.
arXiv Detail & Related papers (2022-12-18T02:03:00Z) - A Self-Attention Ansatz for Ab-initio Quantum Chemistry [3.4161707164978137]
We present a novel neural network architecture using self-attention, the Wavefunction Transformer (Psiformer)
We show that the Psiformer can be used as a drop-in replacement for other neural networks, often dramatically improving the accuracy of the calculations.
This demonstrates that self-attention networks can learn complex quantum mechanical correlations between electrons, and are a promising route to reaching unprecedented accuracy in chemical calculations on larger systems.
arXiv Detail & Related papers (2022-11-24T15:38:55Z) - Physics-informed machine learning with differentiable programming for
heterogeneous underground reservoir pressure management [64.17887333976593]
Avoiding over-pressurization in subsurface reservoirs is critical for applications like CO2 sequestration and wastewater injection.
Managing the pressures by controlling injection/extraction are challenging because of complex heterogeneity in the subsurface.
We use differentiable programming with a full-physics model and machine learning to determine the fluid extraction rates that prevent over-pressurization.
arXiv Detail & Related papers (2022-06-21T20:38:13Z) - Equivariant Graph Mechanics Networks with Constraints [83.38709956935095]
We propose Graph Mechanics Network (GMN) which is efficient, equivariant and constraint-aware.
GMN represents, by generalized coordinates, the forward kinematics information (positions and velocities) of a structural object.
Extensive experiments support the advantages of GMN compared to the state-of-the-art GNNs in terms of prediction accuracy, constraint satisfaction and data efficiency.
arXiv Detail & Related papers (2022-03-12T14:22:14Z) - Physics Informed RNN-DCT Networks for Time-Dependent Partial
Differential Equations [62.81701992551728]
We present a physics-informed framework for solving time-dependent partial differential equations.
Our model utilizes discrete cosine transforms to encode spatial and recurrent neural networks.
We show experimental results on the Taylor-Green vortex solution to the Navier-Stokes equations.
arXiv Detail & Related papers (2022-02-24T20:46:52Z) - Physics-Guided Problem Decomposition for Scaling Deep Learning of
High-dimensional Eigen-Solvers: The Case of Schr\"{o}dinger's Equation [8.80823317679047]
Deep neural networks (NNs) have been proposed as a viable alternative to traditional simulation-driven approaches for solving high-dimensional eigenvalue equations.
In this paper, we use physics knowledge to decompose the complex regression task of predicting the high-dimensional eigenvectors into simpler sub-tasks.
We demonstrate the efficacy of such physics-guided problem decomposition for the case of the Schr"odinger's Equation in Quantum Mechanics.
arXiv Detail & Related papers (2022-02-12T05:59:08Z) - Characterizing possible failure modes in physics-informed neural
networks [55.83255669840384]
Recent work in scientific machine learning has developed so-called physics-informed neural network (PINN) models.
We demonstrate that, while existing PINN methodologies can learn good models for relatively trivial problems, they can easily fail to learn relevant physical phenomena even for simple PDEs.
We show that these possible failure modes are not due to the lack of expressivity in the NN architecture, but that the PINN's setup makes the loss landscape very hard to optimize.
arXiv Detail & Related papers (2021-09-02T16:06:45Z) - Physics-Integrated Variational Autoencoders for Robust and Interpretable
Generative Modeling [86.9726984929758]
We focus on the integration of incomplete physics models into deep generative models.
We propose a VAE architecture in which a part of the latent space is grounded by physics.
We demonstrate generative performance improvements over a set of synthetic and real-world datasets.
arXiv Detail & Related papers (2021-02-25T20:28:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.