Unsupervised Pool-Based Active Learning for Linear Regression
- URL: http://arxiv.org/abs/2001.05028v1
- Date: Tue, 14 Jan 2020 20:00:10 GMT
- Title: Unsupervised Pool-Based Active Learning for Linear Regression
- Authors: Ziang Liu and Dongrui Wu
- Abstract summary: This paper studies unsupervised pool-based AL for linear regression problems.
We propose a novel AL approach that considers simultaneously the informativeness, representativeness, and diversity, three essential criteria in AL.
- Score: 29.321275647107928
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many real-world machine learning applications, unlabeled data can be
easily obtained, but it is very time-consuming and/or expensive to label them.
So, it is desirable to be able to select the optimal samples to label, so that
a good machine learning model can be trained from a minimum amount of labeled
data. Active learning (AL) has been widely used for this purpose. However, most
existing AL approaches are supervised: they train an initial model from a small
amount of labeled samples, query new samples based on the model, and then
update the model iteratively. Few of them have considered the completely
unsupervised AL problem, i.e., starting from zero, how to optimally select the
very first few samples to label, without knowing any label information at all.
This problem is very challenging, as no label information can be utilized. This
paper studies unsupervised pool-based AL for linear regression problems. We
propose a novel AL approach that considers simultaneously the informativeness,
representativeness, and diversity, three essential criteria in AL. Extensive
experiments on 14 datasets from various application domains, using three
different linear regression models (ridge regression, LASSO, and linear support
vector regression), demonstrated the effectiveness of our proposed approach.
Related papers
- One-shot Active Learning Based on Lewis Weight Sampling for Multiple Deep Models [39.582100727546816]
Active learning (AL) for multiple target models aims to reduce labeled data querying while effectively training multiple models concurrently.
Existing AL algorithms often rely on iterative model training, which can be computationally expensive.
We propose a one-shot AL method to address this challenge, which performs all label queries without repeated model training.
arXiv Detail & Related papers (2024-05-23T02:48:16Z) - Automated Labeling of German Chest X-Ray Radiology Reports using Deep
Learning [50.591267188664666]
We propose a deep learning-based CheXpert label prediction model, pre-trained on reports labeled by a rule-based German CheXpert model.
Our results demonstrate the effectiveness of our approach, which significantly outperformed the rule-based model on all three tasks.
arXiv Detail & Related papers (2023-06-09T16:08:35Z) - Label-Retrieval-Augmented Diffusion Models for Learning from Noisy
Labels [61.97359362447732]
Learning from noisy labels is an important and long-standing problem in machine learning for real applications.
In this paper, we reformulate the label-noise problem from a generative-model perspective.
Our model achieves new state-of-the-art (SOTA) results on all the standard real-world benchmark datasets.
arXiv Detail & Related papers (2023-05-31T03:01:36Z) - Learned Label Aggregation for Weak Supervision [8.819582879892762]
We propose a data programming approach that aggregates weak supervision signals to generate labeled data easily.
The quality of the generated labels depends on a label aggregation model that aggregates all noisy labels from all LFs to infer the ground-truth labels.
We show the model can be trained using synthetically generated data and design an effective architecture for the model.
arXiv Detail & Related papers (2022-07-27T14:36:35Z) - Pareto Optimization for Active Learning under Out-of-Distribution Data
Scenarios [79.02009938011447]
We propose a sampling scheme, which selects optimal subsets of unlabeled samples with fixed batch size from the unlabeled data pool.
Experimental results show its effectiveness on both classical Machine Learning (ML) and Deep Learning (DL) tasks.
arXiv Detail & Related papers (2022-07-04T04:11:44Z) - A Lagrangian Duality Approach to Active Learning [119.36233726867992]
We consider the batch active learning problem, where only a subset of the training data is labeled.
We formulate the learning problem using constrained optimization, where each constraint bounds the performance of the model on labeled samples.
We show, via numerical experiments, that our proposed approach performs similarly to or better than state-of-the-art active learning methods.
arXiv Detail & Related papers (2022-02-08T19:18:49Z) - Improving Contrastive Learning on Imbalanced Seed Data via Open-World
Sampling [96.8742582581744]
We present an open-world unlabeled data sampling framework called Model-Aware K-center (MAK)
MAK follows three simple principles: tailness, proximity, and diversity.
We demonstrate that MAK can consistently improve both the overall representation quality and the class balancedness of the learned features.
arXiv Detail & Related papers (2021-11-01T15:09:41Z) - DEAL: Deep Evidential Active Learning for Image Classification [0.0]
Active Learning (AL) is one approach to mitigate the problem of limited labeled data.
Recent AL methods for CNNs propose different solutions for the selection of instances to be labeled.
We propose a novel AL algorithm that efficiently learns from unlabeled data by capturing high prediction uncertainty.
arXiv Detail & Related papers (2020-07-22T11:14:23Z) - Integrating Informativeness, Representativeness and Diversity in
Pool-Based Sequential Active Learning for Regression [29.321275647107928]
It optimally selects the best few samples to label, so that a better machine learning model can be trained from the same number of labeled samples.
Three essential criteria -- informativeness, representativeness, and diversity -- have been proposed for regression problems.
We propose three new ALR approaches, with different strategies for integrating the three criteria.
arXiv Detail & Related papers (2020-03-26T08:10:58Z) - Pool-Based Unsupervised Active Learning for Regression Using Iterative
Representativeness-Diversity Maximization (iRDM) [22.037639625586667]
Active learning (AL) selects the most beneficial unlabeled samples to label, and hence a better machine learning model can be trained from the same number of labeled samples.
Most existing active learning for regression (ALR) approaches are supervised, which means the sampling process must use some label information.
We propose a novel unsupervised ALR approach, iterative representativeness-diversity (iRDM) to balance the representativeness and the diversity of the selected samples.
arXiv Detail & Related papers (2020-03-17T12:20:46Z) - Progressive Identification of True Labels for Partial-Label Learning [112.94467491335611]
Partial-label learning (PLL) is a typical weakly supervised learning problem, where each training instance is equipped with a set of candidate labels among which only one is the true label.
Most existing methods elaborately designed as constrained optimizations that must be solved in specific manners, making their computational complexity a bottleneck for scaling up to big data.
This paper proposes a novel framework of classifier with flexibility on the model and optimization algorithm.
arXiv Detail & Related papers (2020-02-19T08:35:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.