A Study of Acquisition Functions for Medical Imaging Deep Active
Learning
- URL: http://arxiv.org/abs/2401.15721v2
- Date: Thu, 29 Feb 2024 11:02:02 GMT
- Title: A Study of Acquisition Functions for Medical Imaging Deep Active
Learning
- Authors: Bonaventure F. P. Dossou
- Abstract summary: We show how active learning could be very effective in data scarcity situations.
We compare several selection criteria (BALD, MeanSTD, and MaxEntropy) on the ISIC 2016 dataset.
Our results suggest that uncertainty is useful to the Melanoma detection task.
- Score: 2.4654745083407175
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Deep Learning revolution has enabled groundbreaking achievements in
recent years. From breast cancer detection to protein folding, deep learning
algorithms have been at the core of very important advancements. However, these
modern advancements are becoming more and more data-hungry, especially on
labeled data whose availability is scarce: this is even more prevalent in the
medical context. In this work, we show how active learning could be very
effective in data scarcity situations, where obtaining labeled data (or
annotation budget is very limited). We compare several selection criteria
(BALD, MeanSTD, and MaxEntropy) on the ISIC 2016 dataset. We also explored the
effect of acquired pool size on the model's performance. Our results suggest
that uncertainty is useful to the Melanoma detection task, and confirms the
hypotheses of the author of the paper of interest, that \textit{bald} performs
on average better than other acquisition functions. Our extended analyses
however revealed that all acquisition functions perform badly on the positive
(cancerous) samples, suggesting exploitation of class unbalance, which could be
crucial in real-world settings. We finish by suggesting future work directions
that would be useful to improve this current work. The code of our
implementation is open-sourced at
\url{https://github.com/bonaventuredossou/ece526_course_project}
Related papers
- Generative Active Learning for Long-tailed Instance Segmentation [55.66158205855948]
We propose BSGAL, a new algorithm that estimates the contribution of generated data based on cache gradient.
Experiments show that BSGAL outperforms the baseline approach and effectually improves the performance of long-tailed segmentation.
arXiv Detail & Related papers (2024-06-04T15:57:43Z) - Model Uncertainty based Active Learning on Tabular Data using Boosted
Trees [0.4667030429896303]
Supervised machine learning relies on the availability of good labelled data for model training.
Active learning is a sub-field of machine learning which helps in obtaining the labelled data efficiently.
arXiv Detail & Related papers (2023-10-30T14:29:53Z) - On the Limitations of Simulating Active Learning [32.34440406689871]
Active learning (AL) is a human-and-model-in-the-loop paradigm that iteratively selects informative unlabeled data for human annotation.
An easy fix to this impediment is to simulate AL, by treating an already labeled and publicly available dataset as the pool of unlabeled data.
We argue that evaluating AL algorithms on available labeled datasets might provide a lower bound as to their effectiveness in real data.
arXiv Detail & Related papers (2023-05-21T22:52:13Z) - Reinforcement Learning from Passive Data via Latent Intentions [86.4969514480008]
We show that passive data can still be used to learn features that accelerate downstream RL.
Our approach learns from passive data by modeling intentions.
Our experiments demonstrate the ability to learn from many forms of passive data, including cross-embodiment video data and YouTube videos.
arXiv Detail & Related papers (2023-04-10T17:59:05Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Adversarial Auto-Augment with Label Preservation: A Representation
Learning Principle Guided Approach [95.74102207187545]
We show that a prior-free autonomous data augmentation's objective can be derived from a representation learning principle.
We then propose a practical surrogate to the objective that can be efficiently optimized and integrated seamlessly into existing methods.
arXiv Detail & Related papers (2022-11-02T02:02:51Z) - Performance of different machine learning methods on activity
recognition and pose estimation datasets [0.0]
This paper employs both classical and ensemble approaches on rich pose estimation (OpenPose) and HAR datasets.
The results show that overall, random forest yields the highest accuracy in classifying ADLs.
Relatively all the models have excellent performance across both datasets, except for logistic regression and AdaBoost perform poorly in the HAR one.
arXiv Detail & Related papers (2022-10-19T02:07:43Z) - From Weakly Supervised Learning to Active Learning [1.52292571922932]
This thesis is motivated by the question: can we derive a more generic framework than the one of supervised learning?
We model weak supervision as giving, rather than a unique target, a set of target candidates.
We argue that one should look for an optimistic'' function that matches most of the observations. This allows us to derive a principle to disambiguate partial labels.
arXiv Detail & Related papers (2022-09-23T14:55:43Z) - Bayesian Active Learning for Wearable Stress and Affect Detection [0.7106986689736827]
Stress detection using on-device deep learning algorithms has been on the rise owing to advancements in pervasive computing.
In this paper, we propose a framework with capabilities to represent model uncertainties through approximations in Bayesian Neural Networks.
Our proposed framework achieves a considerable efficiency boost during inference, with a substantially low number of acquired pool points.
arXiv Detail & Related papers (2020-12-04T16:19:37Z) - Bayesian active learning for production, a systematic study and a
reusable library [85.32971950095742]
In this paper, we analyse the main drawbacks of current active learning techniques.
We do a systematic study on the effects of the most common issues of real-world datasets on the deep active learning process.
We derive two techniques that can speed up the active learning loop such as partial uncertainty sampling and larger query size.
arXiv Detail & Related papers (2020-06-17T14:51:11Z) - Confident Coreset for Active Learning in Medical Image Analysis [57.436224561482966]
We propose a novel active learning method, confident coreset, which considers both uncertainty and distribution for effectively selecting informative samples.
By comparative experiments on two medical image analysis tasks, we show that our method outperforms other active learning methods.
arXiv Detail & Related papers (2020-04-05T13:46:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.