Uncertainty Prediction for Machine Learning Models of Material
Properties
- URL: http://arxiv.org/abs/2107.07997v1
- Date: Fri, 16 Jul 2021 16:33:55 GMT
- Title: Uncertainty Prediction for Machine Learning Models of Material
Properties
- Authors: Francesca Tavazza, Brian De Cost, Kamal Choudhary
- Abstract summary: Uncertainty in AI-based predictions of material properties is of immense importance for the success and reliability of AI applications in material science.
We compare 3 different approaches to obtain such individual uncertainty, testing them on 12 ML-physical properties.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Uncertainty quantification in Artificial Intelligence (AI)-based predictions
of material properties is of immense importance for the success and reliability
of AI applications in material science. While confidence intervals are commonly
reported for machine learning (ML) models, prediction intervals, i.e., the
evaluation of the uncertainty on each prediction, are seldomly available. In
this work we compare 3 different approaches to obtain such individual
uncertainty, testing them on 12 ML-physical properties. Specifically, we
investigated using the Quantile loss function, machine learning the prediction
intervals directly and using Gaussian Processes. We identify each approachs
advantages and disadvantages and end up slightly favoring the modeling of the
individual uncertainties directly, as it is the easiest to fit and, in most
cases, minimizes over-and under-estimation of the predicted errors. All data
for training and testing were taken from the publicly available JARVIS-DFT
database, and the codes developed for computing the prediction intervals are
available through JARVIS-Tools.
Related papers
- Comprehensive Study Of Predictive Maintenance In Industries Using Classification Models And LSTM Model [0.0]
The study aims to delve into various machine learning classification techniques, including Support Vector Machine (SVM), Random Forest, Logistic Regression, and Convolutional Neural Network LSTM-Based, for predicting and analyzing machine performance.
The primary objective of the study is to assess these algorithms' performance in predicting and analyzing machine performance, considering factors such as accuracy, precision, recall, and F1 score.
arXiv Detail & Related papers (2024-03-15T12:47:45Z) - Variance of ML-based software fault predictors: are we really improving
fault prediction? [0.3222802562733786]
We experimentally analyze the variance of a state-of-the-art fault prediction approach.
We observed a maximum variance of 10.10% in terms of the per-class accuracy metric.
arXiv Detail & Related papers (2023-10-26T09:31:32Z) - Approaches for Uncertainty Quantification of AI-predicted Material
Properties: A Comparison [0.4037357056611557]
Three easy-to-implement approaches to determine individual uncertainty are presented.
We focus on the direct machine learning of the prediction intervals and Ensemble methods.
arXiv Detail & Related papers (2023-10-19T20:20:39Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - On the Relation between Prediction and Imputation Accuracy under Missing
Covariates [0.0]
Recent research has realized an increasing trend towards the usage of modern Machine Learning algorithms for imputation.
Recent research has realized an increasing trend towards the usage of modern Machine Learning algorithms for imputation.
arXiv Detail & Related papers (2021-12-09T23:30:44Z) - LIFE: Learning Individual Features for Multivariate Time Series
Prediction with Missing Values [71.52335136040664]
We propose a Learning Individual Features (LIFE) framework, which provides a new paradigm for MTS prediction with missing values.
LIFE generates reliable features for prediction by using the correlated dimensions as auxiliary information and suppressing the interference from uncorrelated dimensions with missing values.
Experiments on three real-world data sets verify the superiority of LIFE to existing state-of-the-art models.
arXiv Detail & Related papers (2021-09-30T04:53:24Z) - Hessian-based toolbox for reliable and interpretable machine learning in
physics [58.720142291102135]
We present a toolbox for interpretability and reliability, extrapolation of the model architecture.
It provides a notion of the influence of the input data on the prediction at a given test point, an estimation of the uncertainty of the model predictions, and an agnostic score for the model predictions.
Our work opens the road to the systematic use of interpretability and reliability methods in ML applied to physics and, more generally, science.
arXiv Detail & Related papers (2021-08-04T16:32:59Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - A framework for predicting, interpreting, and improving Learning
Outcomes [0.0]
We develop an Embibe Score Quotient model (ESQ) to predict test scores based on observed academic, behavioral and test-taking features of a student.
ESQ can be used to predict the future scoring potential of a student as well as offer personalized learning nudges.
arXiv Detail & Related papers (2020-10-06T11:22:27Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.