A probabilistic generative model for semi-supervised training of
coarse-grained surrogates and enforcing physical constraints through virtual
observables
- URL: http://arxiv.org/abs/2006.01789v1
- Date: Tue, 2 Jun 2020 17:14:36 GMT
- Title: A probabilistic generative model for semi-supervised training of
coarse-grained surrogates and enforcing physical constraints through virtual
observables
- Authors: Maximilian Rixner, Phaedon-Stelios Koutsourelakis
- Abstract summary: This paper provides a flexible, probabilistic framework that accounts for physical structure and information both in the training objectives and in the surrogate model itself.
We advocate a probabilistic model in which equalities that are available from the physics can be introduced as virtual observables and can provide additional information through the likelihood.
- Score: 3.8073142980733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The data-centric construction of inexpensive surrogates for fine-grained,
physical models has been at the forefront of computational physics due to its
significant utility in many-query tasks such as uncertainty quantification.
Recent efforts have taken advantage of the enabling technologies from the field
of machine learning (e.g. deep neural networks) in combination with simulation
data. While such strategies have shown promise even in higher-dimensional
problems, they generally require large amounts of training data even though the
construction of surrogates is by definition a Small Data problem. Rather than
employing data-based loss functions, it has been proposed to make use of the
governing equations (in the simplest case at collocation points) in order to
imbue domain knowledge in the training of the otherwise black-box-like
interpolators. The present paper provides a flexible, probabilistic framework
that accounts for physical structure and information both in the training
objectives as well as in the surrogate model itself. We advocate a
probabilistic (Bayesian) model in which equalities that are available from the
physics (e.g. residuals, conservation laws) can be introduced as virtual
observables and can provide additional information through the likelihood. We
further advocate a generative model i.e. one that attempts to learn the joint
density of inputs and outputs that is capable of making use of unlabeled data
(i.e. only inputs) in a semi-supervised fashion in order to promote the
discovery of lower-dimensional embeddings which are nevertheless predictive of
the fine-grained model's output.
Related papers
- Optimal design of experiments in the context of machine-learning inter-atomic potentials: improving the efficiency and transferability of kernel based methods [0.7234862895932991]
Data-driven, machine learning (ML) models of atomistic interactions can relate nuanced aspects of atomic arrangements into predictions of energies and forces.
The main challenge stems from the fact that descriptors of chemical environments are often sparse high-dimensional objects without a well-defined continuous metric.
We will demonstrate that classical concepts of statistical planning of experiments and optimal design can help to mitigate such problems at a relatively low computational cost.
arXiv Detail & Related papers (2024-05-14T14:14:23Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - A spectrum of physics-informed Gaussian processes for regression in
engineering [0.0]
Despite the growing availability of sensing and data in general, we remain unable to fully characterise many in-service engineering systems and structures from a purely data-driven approach.
This paper pursues the combination of machine learning technology and physics-based reasoning to enhance our ability to make predictive models with limited data.
arXiv Detail & Related papers (2023-09-19T14:39:03Z) - Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - Physics-aware, deep probabilistic modeling of multiscale dynamics in the
Small Data regime [0.0]
The present paper offers a probabilistic perspective that simultaneously identifies predictive, lower-dimensional coarse-grained (CG) variables as well as their dynamics.
We make use of the expressive ability of deep neural networks in order to represent the right-hand side of the CG evolution law.
We demonstrate the efficacy of the proposed framework in a high-dimensional system of moving particles.
arXiv Detail & Related papers (2021-02-08T15:04:05Z) - Generative Counterfactuals for Neural Networks via Attribute-Informed
Perturbation [51.29486247405601]
We design a framework to generate counterfactuals for raw data instances with the proposed Attribute-Informed Perturbation (AIP)
By utilizing generative models conditioned with different attributes, counterfactuals with desired labels can be obtained effectively and efficiently.
Experimental results on real-world texts and images demonstrate the effectiveness, sample quality as well as efficiency of our designed framework.
arXiv Detail & Related papers (2021-01-18T08:37:13Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - Embedded-physics machine learning for coarse-graining and collective
variable discovery without data [3.222802562733787]
We present a novel learning framework that consistently embeds underlying physics.
We propose a novel objective based on reverse Kullback-Leibler divergence that fully incorporates the available physics in the form of the atomistic force field.
We demonstrate the algorithmic advances in terms of predictive ability and the physical meaning of the revealed CVs for a bimodal potential energy function and the alanine dipeptide.
arXiv Detail & Related papers (2020-02-24T10:28:41Z) - Incorporating physical constraints in a deep probabilistic machine
learning framework for coarse-graining dynamical systems [7.6146285961466]
This paper offers a data-based, probablistic perspective that enables the quantification of predictive uncertainties.
We formulate the coarse-graining process by employing a probabilistic state-space model.
It is capable of reconstructing the evolution of the full, fine-scale system.
arXiv Detail & Related papers (2019-12-30T16:07:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.