Probabilistic Neural Data Fusion for Learning from an Arbitrary Number
of Multi-fidelity Data Sets
- URL: http://arxiv.org/abs/2301.13271v1
- Date: Mon, 30 Jan 2023 20:27:55 GMT
- Title: Probabilistic Neural Data Fusion for Learning from an Arbitrary Number
of Multi-fidelity Data Sets
- Authors: Carlos Mora, Jonathan Tammer Eweis-Labolle, Tyler Johnson, Likith
Gadde, Ramin Bostanabad
- Abstract summary: In this paper, we employ neural networks (NNs) for data fusion in scenarios where data is very scarce.
We introduce a unique NN architecture that converts MF modeling into a nonlinear manifold learning problem.
Our approach provides a high predictive power while quantifying various sources uncertainties.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many applications in engineering and sciences analysts have simultaneous
access to multiple data sources. In such cases, the overall cost of acquiring
information can be reduced via data fusion or multi-fidelity (MF) modeling
where one leverages inexpensive low-fidelity (LF) sources to reduce the
reliance on expensive high-fidelity (HF) data. In this paper, we employ neural
networks (NNs) for data fusion in scenarios where data is very scarce and
obtained from an arbitrary number of sources with varying levels of fidelity
and cost. We introduce a unique NN architecture that converts MF modeling into
a nonlinear manifold learning problem. Our NN architecture inversely learns
non-trivial (e.g., non-additive and non-hierarchical) biases of the LF sources
in an interpretable and visualizable manifold where each data source is encoded
via a low-dimensional distribution. This probabilistic manifold quantifies
model form uncertainties such that LF sources with small bias are encoded close
to the HF source. Additionally, we endow the output of our NN with a parametric
distribution not only to quantify aleatoric uncertainties, but also to
reformulate the network's loss function based on strictly proper scoring rules
which improve robustness and accuracy on unseen HF data. Through a set of
analytic and engineering examples, we demonstrate that our approach provides a
high predictive power while quantifying various sources uncertainties.
Related papers
- Multi-Fidelity Bayesian Neural Network for Uncertainty Quantification in Transonic Aerodynamic Loads [0.0]
This paper implements a multi-fidelity Bayesian neural network model that applies transfer learning to fuse data generated by models at different fidelities.
The results demonstrate that the multi-fidelity Bayesian model outperforms the state-of-the-art Co-Kriging in terms of overall accuracy and robustness on unseen data.
arXiv Detail & Related papers (2024-07-08T07:34:35Z) - Multi-Fidelity Residual Neural Processes for Scalable Surrogate Modeling [19.60087366873302]
Multi-fidelity surrogate modeling aims to learn an accurate surrogate at the highest fidelity level.
Deep learning approaches utilize neural network based encoders and decoders to improve scalability.
We propose Multi-fidelity Residual Neural Processes (MFRNP), a novel multi-fidelity surrogate modeling framework.
arXiv Detail & Related papers (2024-02-29T04:40:25Z) - Interpretable Multi-Source Data Fusion Through Latent Variable Gaussian Process [8.207427766052044]
The proposed approach is demonstrated on and analyzed through two mathematical (representative parabola problem, 2D Ackley function) and two materials science (design of FeCrAl and SmCoFe alloys) case studies.
It is observed that compared to using single-source and source unaware ML models, the proposed multi-source data fusion framework can provide better predictions for sparse-data problems, interpretability regarding the sources, and enhanced modeling capabilities by taking advantage of the correlations and relationships among different sources.
arXiv Detail & Related papers (2024-02-06T16:54:59Z) - Analysis and Optimization of Wireless Federated Learning with Data
Heterogeneity [72.85248553787538]
This paper focuses on performance analysis and optimization for wireless FL, considering data heterogeneity, combined with wireless resource allocation.
We formulate the loss function minimization problem, under constraints on long-term energy consumption and latency, and jointly optimize client scheduling, resource allocation, and the number of local training epochs (CRE)
Experiments on real-world datasets demonstrate that the proposed algorithm outperforms other benchmarks in terms of the learning accuracy and energy consumption.
arXiv Detail & Related papers (2023-08-04T04:18:01Z) - Uncertainty-guided Source-free Domain Adaptation [77.3844160723014]
Source-free domain adaptation (SFDA) aims to adapt a classifier to an unlabelled target data set by only using a pre-trained source model.
We propose quantifying the uncertainty in the source model predictions and utilizing it to guide the target adaptation.
arXiv Detail & Related papers (2022-08-16T08:03:30Z) - Re-parameterizing VAEs for stability [1.90365714903665]
We propose a theoretical approach towards the training numerical stability of Variational AutoEncoders (VAE)
Our work is motivated by recent studies empowering VAEs to reach state of the art generative results on complex image datasets.
We show that by implementing small changes to the way we parameterize the Normal distributions on which they rely, VAEs can securely be trained.
arXiv Detail & Related papers (2021-06-25T16:19:09Z) - Unsupervised Multi-source Domain Adaptation Without Access to Source
Data [58.551861130011886]
Unsupervised Domain Adaptation (UDA) aims to learn a predictor model for an unlabeled domain by transferring knowledge from a separate labeled source domain.
We propose a novel and efficient algorithm which automatically combines the source models with suitable weights in such a way that it performs at least as good as the best source model.
arXiv Detail & Related papers (2021-04-05T10:45:12Z) - MFNets: Data efficient all-at-once learning of multifidelity surrogates
as directed networks of information sources [0.0]
We present an approach for constructing a surrogate from ensembles of information sources of varying cost and accuracy.
The multifidelity surrogate encodes connections between information sources as a directed acyclic graph, and is trained via gradient-based minimization of a nonlinear least squares objective.
arXiv Detail & Related papers (2020-08-03T16:21:33Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.