Robust Modeling of Unknown Dynamical Systems via Ensemble Averaged
Learning
- URL: http://arxiv.org/abs/2203.03458v1
- Date: Mon, 7 Mar 2022 15:17:53 GMT
- Title: Robust Modeling of Unknown Dynamical Systems via Ensemble Averaged
Learning
- Authors: Victor Churchill, Steve Manns, Zhen Chen, Dongbin Xiu
- Abstract summary: Recent work has focused on data-driven learning of the evolution of unknown systems via deep neural networks (DNNs)
This paper presents a computational technique which decreases the variance of the generalization error.
- Score: 2.523610673302386
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work has focused on data-driven learning of the evolution of unknown
systems via deep neural networks (DNNs), with the goal of conducting long time
prediction of the evolution of the unknown system. Training a DNN with low
generalization error is a particularly important task in this case as error is
accumulated over time. Because of the inherent randomness in DNN training,
chiefly in stochastic optimization, there is uncertainty in the resulting
prediction, and therefore in the generalization error. Hence, the
generalization error can be viewed as a random variable with some probability
distribution. Well-trained DNNs, particularly those with many hyperparameters,
typically result in probability distributions for generalization error with low
bias but high variance. High variance causes variability and unpredictably in
the results of a trained DNN. This paper presents a computational technique
which decreases the variance of the generalization error, thereby improving the
reliability of the DNN model to generalize consistently. In the proposed
ensemble averaging method, multiple models are independently trained and model
predictions are averaged at each time step. A mathematical foundation for the
method is presented, including results regarding the distribution of the local
truncation error. In addition, three time-dependent differential equation
problems are considered as numerical examples, demonstrating the effectiveness
of the method to decrease variance of DNN predictions generally.
Related papers
- Generalization of Graph Neural Networks is Robust to Model Mismatch [84.01980526069075]
Graph neural networks (GNNs) have demonstrated their effectiveness in various tasks supported by their generalization capabilities.
In this paper, we examine GNNs that operate on geometric graphs generated from manifold models.
Our analysis reveals the robustness of the GNN generalization in the presence of such model mismatch.
arXiv Detail & Related papers (2024-08-25T16:00:44Z) - Deep Limit Model-free Prediction in Regression [0.0]
We provide a Model-free approach based on Deep Neural Network (DNN) to accomplish point prediction and prediction interval under a general regression setting.
Our method is more stable and accurate compared to other DNN-based counterparts, especially for optimal point predictions.
arXiv Detail & Related papers (2024-08-18T16:37:53Z) - Graph Out-of-Distribution Generalization via Causal Intervention [69.70137479660113]
We introduce a conceptually simple yet principled approach for training robust graph neural networks (GNNs) under node-level distribution shifts.
Our method resorts to a new learning objective derived from causal inference that coordinates an environment estimator and a mixture-of-expert GNN predictor.
Our model can effectively enhance generalization with various types of distribution shifts and yield up to 27.4% accuracy improvement over state-of-the-arts on graph OOD generalization benchmarks.
arXiv Detail & Related papers (2024-02-18T07:49:22Z) - Discovering Invariant Rationales for Graph Neural Networks [104.61908788639052]
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features.
We propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs.
arXiv Detail & Related papers (2022-01-30T16:43:40Z) - Optimization Variance: Exploring Generalization Properties of DNNs [83.78477167211315]
The test error of a deep neural network (DNN) often demonstrates double descent.
We propose a novel metric, optimization variance (OV), to measure the diversity of model updates.
arXiv Detail & Related papers (2021-06-03T09:34:17Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Memorizing without overfitting: Bias, variance, and interpolation in
over-parameterized models [0.0]
The bias-variance trade-off is a central concept in supervised learning.
Modern Deep Learning methods flout this dogma, achieving state-of-the-art performance.
arXiv Detail & Related papers (2020-10-26T22:31:04Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Rethinking Bias-Variance Trade-off for Generalization of Neural Networks [40.04927952870877]
We provide a simple explanation for this by measuring the bias and variance of neural networks.
We find that variance unimodality occurs robustly for all models we considered.
Deeper models decrease bias and increase variance for both in-distribution and out-of-distribution data.
arXiv Detail & Related papers (2020-02-26T07:21:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.