A comprehensive study on the prediction reliability of graph neural
networks for virtual screening
- URL: http://arxiv.org/abs/2003.07611v1
- Date: Tue, 17 Mar 2020 10:13:31 GMT
- Title: A comprehensive study on the prediction reliability of graph neural
networks for virtual screening
- Authors: Soojung Yang, Kyung Hoon Lee, and Seongok Ryu
- Abstract summary: We investigate the effects of model architectures, regularization methods, and loss functions on the prediction performance and reliability of classification results.
Our result highlights that correct choice of regularization and inference methods is evidently important to achieve high success rate.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prediction models based on deep neural networks are increasingly gaining
attention for fast and accurate virtual screening systems. For decision makings
in virtual screening, researchers find it useful to interpret an output of
classification system as probability, since such interpretation allows them to
filter out more desirable compounds. However, probabilistic interpretation
cannot be correct for models that hold over-parameterization problems or
inappropriate regularizations, leading to unreliable prediction and decision
making. In this regard, we concern the reliability of neural prediction models
on molecular properties, especially when models are trained with sparse data
points and imbalanced distributions. This work aims to propose guidelines for
training reliable models, we thus provide methodological details and ablation
studies on the following train principles. We investigate the effects of model
architectures, regularization methods, and loss functions on the prediction
performance and reliability of classification results. Moreover, we evaluate
prediction reliability of models on virtual screening scenario. Our result
highlights that correct choice of regularization and inference methods is
evidently important to achieve high success rate, especially in data imbalanced
situation. All experiments were performed under a single unified model
implementation to alleviate external randomness in model training and to enable
precise comparison of results.
Related papers
- Detecting Model Misspecification in Amortized Bayesian Inference with Neural Networks: An Extended Investigation [9.950524371154394]
We propose a new misspecification measure that can be trained in an unsupervised fashion and reliably detects model misspecification at test time.
We show how the proposed misspecification test warns users about suspicious outputs, raises an alarm when predictions are not trustworthy, and guides model designers in their search for better simulators.
arXiv Detail & Related papers (2024-06-05T11:30:16Z) - Structured Radial Basis Function Network: Modelling Diversity for
Multiple Hypotheses Prediction [51.82628081279621]
Multi-modal regression is important in forecasting nonstationary processes or with a complex mixture of distributions.
A Structured Radial Basis Function Network is presented as an ensemble of multiple hypotheses predictors for regression problems.
It is proved that this structured model can efficiently interpolate this tessellation and approximate the multiple hypotheses target distribution.
arXiv Detail & Related papers (2023-09-02T01:27:53Z) - Causal Analysis for Robust Interpretability of Neural Networks [0.2519906683279152]
We develop a robust interventional-based method to capture cause-effect mechanisms in pre-trained neural networks.
We apply our method to vision models trained on classification tasks.
arXiv Detail & Related papers (2023-05-15T18:37:24Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Correcting Model Bias with Sparse Implicit Processes [0.9187159782788579]
We show that Sparse Implicit Processes (SIP) is capable of correcting model bias when the data generating mechanism differs strongly from the one implied by the model.
We use synthetic datasets to show that SIP is capable of providing predictive distributions that reflect the data better than the exact predictions of the initial, but wrongly assumed model.
arXiv Detail & Related papers (2022-07-21T18:00:01Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Quantifying Predictive Uncertainty in Medical Image Analysis with Deep
Kernel Learning [14.03923026690186]
We propose an uncertainty-aware deep kernel learning model which permits the estimation of the uncertainty in the prediction.
In most cases, the proposed model shows better performance compared to common architectures.
Our model can also be used to detect challenging and controversial test samples.
arXiv Detail & Related papers (2021-06-01T17:09:47Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Trust but Verify: Assigning Prediction Credibility by Counterfactual
Constrained Learning [123.3472310767721]
Prediction credibility measures are fundamental in statistics and machine learning.
These measures should account for the wide variety of models used in practice.
The framework developed in this work expresses the credibility as a risk-fit trade-off.
arXiv Detail & Related papers (2020-11-24T19:52:38Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.