Probabilistic Attention based on Gaussian Processes for Deep Multiple
Instance Learning
- URL: http://arxiv.org/abs/2302.04061v1
- Date: Wed, 8 Feb 2023 13:58:11 GMT
- Title: Probabilistic Attention based on Gaussian Processes for Deep Multiple
Instance Learning
- Authors: Arne Schmidt, Pablo Morales-\'Alvarez, Rafael Molina
- Abstract summary: We introduce the Attention Gaussian Process (AGP) model, a novel probabilistic attention mechanism based on Gaussian Processes for deep MIL.
AGP provides accurate bag-level predictions as well as instance-level explainability, and can be trained end-to-end.
We experimentally show that predictive uncertainty correlates with the risk of wrong predictions, and therefore it is a good indicator of reliability in practice.
- Score: 12.594098548008832
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiple Instance Learning (MIL) is a weakly supervised learning paradigm
that is becoming increasingly popular because it requires less labeling effort
than fully supervised methods. This is especially interesting for areas where
the creation of large annotated datasets remains challenging, as in medicine.
Although recent deep learning MIL approaches have obtained state-of-the-art
results, they are fully deterministic and do not provide uncertainty
estimations for the predictions. In this work, we introduce the Attention
Gaussian Process (AGP) model, a novel probabilistic attention mechanism based
on Gaussian Processes for deep MIL. AGP provides accurate bag-level predictions
as well as instance-level explainability, and can be trained end-to-end.
Moreover, its probabilistic nature guarantees robustness to overfitting on
small datasets and uncertainty estimations for the predictions. The latter is
especially important in medical applications, where decisions have a direct
impact on the patient's health. The proposed model is validated experimentally
as follows. First, its behavior is illustrated in two synthetic MIL experiments
based on the well-known MNIST and CIFAR-10 datasets, respectively. Then, it is
evaluated in three different real-world cancer detection experiments. AGP
outperforms state-of-the-art MIL approaches, including deterministic deep
learning ones. It shows a strong performance even on a small dataset with less
than 100 labels and generalizes better than competing methods on an external
test set. Moreover, we experimentally show that predictive uncertainty
correlates with the risk of wrong predictions, and therefore it is a good
indicator of reliability in practice. Our code is publicly available.
Related papers
- Evaluation of uncertainty estimations for Gaussian process regression based machine learning interatomic potentials [0.0]
Uncertainty estimations for machine learning interatomic potentials are crucial to quantify the additional model error they introduce.
We consider GPR models with Coulomb and SOAP representations as inputs to predict potential energy surfaces and excitation energies of molecules.
We evaluate, how the GPR variance and ensemble-based uncertainties relate to the error and whether model performance improves by selecting the most uncertain samples from a fixed configuration space.
arXiv Detail & Related papers (2024-10-27T10:06:09Z) - SepsisLab: Early Sepsis Prediction with Uncertainty Quantification and Active Sensing [67.8991481023825]
Sepsis is the leading cause of in-hospital mortality in the USA.
Existing predictive models are usually trained on high-quality data with few missing information.
For the potential high-risk patients with low confidence due to limited observations, we propose a robust active sensing algorithm.
arXiv Detail & Related papers (2024-07-24T04:47:36Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Introducing instance label correlation in multiple instance learning.
Application to cancer detection on histopathological images [5.895585247199983]
In this work, we extend a state-of-the-art GP-based MIL method, which is called VGPMIL-PR, to exploit such correlation.
We show that our model achieves better results than other state-of-the-art probabilistic MIL methods.
arXiv Detail & Related papers (2023-10-30T08:57:59Z) - B-Learner: Quasi-Oracle Bounds on Heterogeneous Causal Effects Under
Hidden Confounding [51.74479522965712]
We propose a meta-learner called the B-Learner, which can efficiently learn sharp bounds on the CATE function under limits on hidden confounding.
We prove its estimates are valid, sharp, efficient, and have a quasi-oracle property with respect to the constituent estimators under more general conditions than existing methods.
arXiv Detail & Related papers (2023-04-20T18:07:19Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - A Benchmark on Uncertainty Quantification for Deep Learning Prognostics [0.0]
We assess some of the latest developments in the field of uncertainty quantification for prognostics deep learning.
This includes the state-of-the-art variational inference algorithms for Bayesian neural networks (BNN) as well as popular alternatives such as Monte Carlo Dropout (MCD), deep ensembles (DE) and heteroscedastic neural networks (HNN)
The performance of the methods is evaluated on a subset of the large NASA NCMAPSS dataset for aircraft engines.
arXiv Detail & Related papers (2023-02-09T16:12:47Z) - Modeling Disagreement in Automatic Data Labelling for Semi-Supervised
Learning in Clinical Natural Language Processing [2.016042047576802]
We investigate the quality of uncertainty estimates from a range of current state-of-the-art predictive models applied to the problem of observation detection in radiology reports.
arXiv Detail & Related papers (2022-05-29T20:20:49Z) - MAg: a simple learning-based patient-level aggregation method for
detecting microsatellite instability from whole-slide images [3.0134189693277]
The prediction of microsatellite instability (MSI) and microsatellite stability (MSS) is essential in predicting both the treatment response and prognosis of gastrointestinal cancer.
Deep-learning-based algorithms have been proposed to predict MSI directly from haematoxylin and eosin (H&E)-stained whole-slide images (WSIs)
We propose a simple yet broadly generalizable patient-level MSI aggregation (MAg) method to effectively integrate the precious patch-level information.
arXiv Detail & Related papers (2022-01-13T02:53:55Z) - Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic
Regression [51.770998056563094]
Probabilistic Gradient Boosting Machines (PGBM) is a method to create probabilistic predictions with a single ensemble of decision trees.
We empirically demonstrate the advantages of PGBM compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-03T08:32:13Z) - UNITE: Uncertainty-based Health Risk Prediction Leveraging Multi-sourced
Data [81.00385374948125]
We present UNcertaInTy-based hEalth risk prediction (UNITE) model.
UNITE provides accurate disease risk prediction and uncertainty estimation leveraging multi-sourced health data.
We evaluate UNITE on real-world disease risk prediction tasks: nonalcoholic fatty liver disease (NASH) and Alzheimer's disease (AD)
UNITE achieves up to 0.841 in F1 score for AD detection, up to 0.609 in PR-AUC for NASH detection, and outperforms various state-of-the-art baselines by up to $19%$ over the best baseline.
arXiv Detail & Related papers (2020-10-22T02:28:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.