Transfer learning suppresses simulation bias in predictive models built
from sparse, multi-modal data
- URL: http://arxiv.org/abs/2104.09684v1
- Date: Mon, 19 Apr 2021 23:28:32 GMT
- Title: Transfer learning suppresses simulation bias in predictive models built
from sparse, multi-modal data
- Authors: Bogdan Kustowski, Jim A. Gaffney, Brian K. Spears, Gemma J. Anderson,
Rushil Anirudh, Peer-Timo Bremer, Jayaraman J. Thiagarajan (Lawrence
Livermore National Laboratory, Livermore, CA)
- Abstract summary: Many problems in science, engineering, and business require making predictions based on very few observations.
To build a robust predictive model, these sparse data may need to be augmented with simulated data, especially when the design space is multidimensional.
We combine recent developments in deep learning to build more robust predictive models from multimodal data.
- Score: 15.587831925516957
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Many problems in science, engineering, and business require making
predictions based on very few observations. To build a robust predictive model,
these sparse data may need to be augmented with simulated data, especially when
the design space is multidimensional. Simulations, however, often suffer from
an inherent bias. Estimation of this bias may be poorly constrained not only
because of data sparsity, but also because traditional predictive models fit
only one type of observations, such as scalars or images, instead of all
available data modalities, which might have been acquired and simulated at
great cost. We combine recent developments in deep learning to build more
robust predictive models from multimodal data with a recent, novel technique to
suppress the bias, and extend it to take into account multiple data modalities.
First, an initial, simulation-trained, neural network surrogate model learns
important correlations between different data modalities and between simulation
inputs and outputs. Then, the model is partially retrained, or transfer
learned, to fit the observations. Using fewer than 10 inertial confinement
fusion experiments for retraining, we demonstrate that this technique
systematically improves simulation predictions while a simple output
calibration makes predictions worse. We also offer extensive cross-validation
with real and synthetic data to support our findings. The transfer learning
method can be applied to other problems that require transferring knowledge
from simulations to the domain of real observations. This paper opens up the
path to model calibration using multiple data types, which have traditionally
been ignored in predictive models.
Related papers
- Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Learning Defect Prediction from Unrealistic Data [57.53586547895278]
Pretrained models of code have become popular choices for code understanding and generation tasks.
Such models tend to be large and require commensurate volumes of training data.
It has become popular to train models with far larger but less realistic datasets, such as functions with artificially injected bugs.
Models trained on such data tend to only perform well on similar data, while underperforming on real world programs.
arXiv Detail & Related papers (2023-11-02T01:51:43Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Learning from aggregated data with a maximum entropy model [73.63512438583375]
We show how a new model, similar to a logistic regression, may be learned from aggregated data only by approximating the unobserved feature distribution with a maximum entropy hypothesis.
We present empirical evidence on several public datasets that the model learned this way can achieve performances comparable to those of a logistic model trained with the full unaggregated data.
arXiv Detail & Related papers (2022-10-05T09:17:27Z) - Cognitive simulation models for inertial confinement fusion: Combining
simulation and experimental data [0.0]
Researchers rely heavily on computer simulations to explore the design space in search of high-performing implosions.
For more effective design and investigation, simulations require input from past experimental data to better predict future performance.
We describe a cognitive simulation method for combining simulation and experimental data into a common, predictive model.
arXiv Detail & Related papers (2021-03-19T02:00:14Z) - Synthesizing Irreproducibility in Deep Networks [2.28438857884398]
Modern day deep networks suffer from irreproducibility (also referred to as nondeterminism or underspecification)
We show that even with a single nonlinearity and for very simple data and models, irreproducibility occurs.
Model complexity and the choice of nonlinearity also play significant roles in making deep models irreproducible.
arXiv Detail & Related papers (2021-02-21T21:51:28Z) - Designing Accurate Emulators for Scientific Processes using
Calibration-Driven Deep Models [33.935755695805724]
Learn-by-Calibrating (LbC) is a novel deep learning approach for designing emulators in scientific applications.
We show that LbC provides significant improvements in generalization error over widely-adopted loss function choices.
LbC achieves high-quality emulators even in small data regimes and more importantly, recovers the inherent noise structure without any explicit priors.
arXiv Detail & Related papers (2020-05-05T16:54:11Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z) - Forecasting Industrial Aging Processes with Machine Learning Methods [0.0]
We evaluate a wider range of data-driven models, comparing some traditional stateless models to more complex recurrent neural networks.
Our results show that recurrent models produce near perfect predictions when trained on larger datasets.
arXiv Detail & Related papers (2020-02-05T13:06:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.