Warm Start Active Learning with Proxy Labels \& Selection via
Semi-Supervised Fine-Tuning
- URL: http://arxiv.org/abs/2209.06285v1
- Date: Tue, 13 Sep 2022 20:21:40 GMT
- Title: Warm Start Active Learning with Proxy Labels \& Selection via
Semi-Supervised Fine-Tuning
- Authors: Vishwesh Nath, Dong Yang, Holger R. Roth, Daguang Xu
- Abstract summary: We propose two novel strategies for active learning (AL) specifically for 3D image segmentation.
First, we tackle the cold start problem by proposing a proxy task and then utilizing uncertainty generated from the proxy task to rank the unlabeled data to be annotated.
Second, we craft a two-stage learning framework for each active iteration where the unlabeled data is also used in the second stage as a semi-supervised fine-tuning strategy.
- Score: 10.086685855244664
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Which volume to annotate next is a challenging problem in building medical
imaging datasets for deep learning. One of the promising methods to approach
this question is active learning (AL). However, AL has been a hard nut to crack
in terms of which AL algorithm and acquisition functions are most useful for
which datasets. Also, the problem is exacerbated with which volumes to label
first when there is zero labeled data to start with. This is known as the cold
start problem in AL. We propose two novel strategies for AL specifically for 3D
image segmentation. First, we tackle the cold start problem by proposing a
proxy task and then utilizing uncertainty generated from the proxy task to rank
the unlabeled data to be annotated. Second, we craft a two-stage learning
framework for each active iteration where the unlabeled data is also used in
the second stage as a semi-supervised fine-tuning strategy. We show the promise
of our approach on two well-known large public datasets from medical
segmentation decathlon. The results indicate that the initial selection of data
and semi-supervised framework both showed significant improvement for several
AL strategies.
Related papers
- Learnability Gaps of Strategic Classification [68.726857356532]
We focus on addressing a fundamental question: the learnability gaps between strategic classification and standard learning.
We provide nearly tight sample complexity and regret bounds, offering significant improvements over prior results.
Notably, our algorithm in this setting is of independent interest and can be applied to other problems such as multi-label learning.
arXiv Detail & Related papers (2024-02-29T16:09:19Z) - MyriadAL: Active Few Shot Learning for Histopathology [10.652626309100889]
We introduce an active few shot learning framework, Myriad Active Learning (MAL)
MAL includes a contrastive-learning encoder, pseudo-label generation, and novel query sample selection in the loop.
Experiments on two public histopathology datasets show that MAL has superior test accuracy, macro F1-score, and label efficiency compared to prior works.
arXiv Detail & Related papers (2023-10-24T20:08:15Z) - Discwise Active Learning for LiDAR Semantic Segmentation [92.92714645987083]
Active learning (AL) provides a solution that can iteratively and intelligently label a dataset.
We propose a discwise approach (DiAL) where in each iteration, we query the region a single frame covers on global coordinates, labeling all frames simultaneously.
arXiv Detail & Related papers (2023-09-23T06:08:22Z) - COLosSAL: A Benchmark for Cold-start Active Learning for 3D Medical
Image Segmentation [10.80144764655265]
Active learning (AL) is a promising solution for efficient annotation but requires an initial set of labeled samples to start active selection.
This is also known as the cold-start AL, which permits only one chance to request annotations from experts without access to previously annotated data.
We present a benchmark named COSAL by evaluating six cold-start AL strategies on five 3D medical segmentation tasks from the public Medical Decathlon collection.
arXiv Detail & Related papers (2023-07-22T07:19:15Z) - On the Limitations of Simulating Active Learning [32.34440406689871]
Active learning (AL) is a human-and-model-in-the-loop paradigm that iteratively selects informative unlabeled data for human annotation.
An easy fix to this impediment is to simulate AL, by treating an already labeled and publicly available dataset as the pool of unlabeled data.
We argue that evaluating AL algorithms on available labeled datasets might provide a lower bound as to their effectiveness in real data.
arXiv Detail & Related papers (2023-05-21T22:52:13Z) - MoBYv2AL: Self-supervised Active Learning for Image Classification [57.4372176671293]
We present MoBYv2AL, a novel self-supervised active learning framework for image classification.
Our contribution lies in lifting MoBY, one of the most successful self-supervised learning algorithms, to the AL pipeline.
We achieve state-of-the-art results when compared to recent AL methods.
arXiv Detail & Related papers (2023-01-04T10:52:02Z) - Deep Active Ensemble Sampling For Image Classification [8.31483061185317]
Active learning frameworks aim to reduce the cost of data annotation by actively requesting the labeling for the most informative data points.
Some proposed approaches include uncertainty-based techniques, geometric methods, implicit combination of uncertainty-based and geometric approaches.
We present an innovative integration of recent progress in both uncertainty-based and geometric frameworks to enable an efficient exploration/exploitation trade-off in sample selection strategy.
Our framework provides two advantages: (1) accurate posterior estimation, and (2) tune-able trade-off between computational overhead and higher accuracy.
arXiv Detail & Related papers (2022-10-11T20:20:20Z) - Communication-Efficient Robust Federated Learning with Noisy Labels [144.31995882209932]
Federated learning (FL) is a promising privacy-preserving machine learning paradigm over distributed located data.
We propose a learning-based reweighting approach to mitigate the effect of noisy labels in FL.
Our approach has shown superior performance on several real-world datasets compared to various baselines.
arXiv Detail & Related papers (2022-06-11T16:21:17Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z) - ATSO: Asynchronous Teacher-Student Optimization for Semi-Supervised
Medical Image Segmentation [99.90263375737362]
We propose ATSO, an asynchronous version of teacher-student optimization.
ATSO partitions the unlabeled data into two subsets and alternately uses one subset to fine-tune the model and updates the label on the other subset.
We evaluate ATSO on two popular medical image segmentation datasets and show its superior performance in various semi-supervised settings.
arXiv Detail & Related papers (2020-06-24T04:05:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.