From Model-driven to Data-driven: A Survey on Active Deep Learning
- URL: http://arxiv.org/abs/2101.09933v1
- Date: Mon, 25 Jan 2021 07:49:41 GMT
- Title: From Model-driven to Data-driven: A Survey on Active Deep Learning
- Authors: Peng Liu, Guojin He, Lei Zhao
- Abstract summary: Active Deep Learning (ADL) only if theirpredictor is deep model, where the basic learner is called as predictor and the labeling schemes iscalled selector.
Wecategory ADL into model-driven ADL and data-driven ADL, by whether its selector is model-driven or data-driven.
The advantages and disadvantages between data-driven ADLand model-driven ADL are thoroughly analyzed.
- Score: 8.75286974962136
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Which samples should be labelled in a large data set is one of the most
important problems for trainingof deep learning. So far, a variety of active
sample selection strategies related to deep learning havebeen proposed in many
literatures. We defined them as Active Deep Learning (ADL) only if
theirpredictor is deep model, where the basic learner is called as predictor
and the labeling schemes iscalled selector. In this survey, three fundamental
factors in selector designation were summarized. Wecategory ADL into
model-driven ADL and data-driven ADL, by whether its selector is model-drivenor
data-driven. The different characteristics of the two major type of ADL were
addressed in indetail respectively. Furthermore, different sub-classes of
data-driven and model-driven ADL are alsosummarized and discussed emphatically.
The advantages and disadvantages between data-driven ADLand model-driven ADL
are thoroughly analyzed. We pointed out that, with the development of
deeplearning, the selector in ADL also is experiencing the stage from
model-driven to data-driven. Finally,we make discussion on ADL about its
uncertainty, explanatory, foundations of cognitive science etc.and survey on
the trend of ADL from model-driven to data-driven.
Related papers
- Automated Detection of Algorithm Debt in Deep Learning Frameworks: An Empirical Study [5.6340045820686155]
Previous studies demonstrate that Machine or Deep Learning (ML/DL) models can detect Technical Debt from source code comments called Self-Admitted Technical Debt (SATD)
Our goal is to improve AD detection performance of various ML/DL models.
arXiv Detail & Related papers (2024-08-20T04:06:58Z) - LESS: Selecting Influential Data for Targeted Instruction Tuning [64.78894228923619]
We propose LESS, an efficient algorithm to estimate data influences and perform Low-rank gradiEnt Similarity Search for instruction data selection.
We show that training on a LESS-selected 5% of the data can often outperform training on the full dataset across diverse downstream tasks.
Our method goes beyond surface form cues to identify data that the necessary reasoning skills for the intended downstream application.
arXiv Detail & Related papers (2024-02-06T19:18:04Z) - The Why, When, and How to Use Active Learning in Large-Data-Driven 3D
Object Detection for Safe Autonomous Driving: An Empirical Exploration [1.2815904071470705]
entropy querying is a promising strategy for selecting data that enhances model learning in resource-constrained environments.
Our findings suggest that entropy querying is a promising strategy for selecting data that enhances model learning in resource-constrained environments.
arXiv Detail & Related papers (2024-01-30T00:14:13Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Self-Evolved Diverse Data Sampling for Efficient Instruction Tuning [47.02160072880698]
We introduce a self-evolving mechanism that allows the model itself to actively sample subsets that are equally or even more effective.
The key to our data sampling technique lies in the enhancement of diversity in the chosen subsets.
Extensive experiments across three datasets and benchmarks demonstrate the effectiveness of DiverseEvol.
arXiv Detail & Related papers (2023-11-14T14:10:40Z) - Zero-Shot AutoML with Pretrained Models [39.928531675761135]
domain-independent meta-learning approach learns a zero-shot surrogate model.
We evaluate our approach under the strict time limit of the ChaLearn AutoDL challenge benchmark.
arXiv Detail & Related papers (2022-06-16T22:52:08Z) - Unified Instance and Knowledge Alignment Pretraining for Aspect-based
Sentiment Analysis [96.53859361560505]
Aspect-based Sentiment Analysis (ABSA) aims to determine the sentiment polarity towards an aspect.
There always exists severe domain shift between the pretraining and downstream ABSA datasets.
We introduce a unified alignment pretraining framework into the vanilla pretrain-finetune pipeline.
arXiv Detail & Related papers (2021-10-26T04:03:45Z) - Knowledge-driven Active Learning [70.37119719069499]
Active learning strategies aim at minimizing the amount of labelled data required to train a Deep Learning model.
Most active strategies are based on uncertain sample selection, and even often restricted to samples lying close to the decision boundary.
Here we propose to take into consideration common domain-knowledge and enable non-expert users to train a model with fewer samples.
arXiv Detail & Related papers (2021-10-15T06:11:53Z) - CorDEL: A Contrastive Deep Learning Approach for Entity Linkage [70.82533554253335]
Entity linkage (EL) is a critical problem in data cleaning and integration.
With the ever-increasing growth of new data, deep learning (DL) based approaches have been proposed to alleviate the high cost of EL associated with the traditional models.
We argue that the twin-network architecture is sub-optimal to EL, leading to inherent drawbacks of existing models.
arXiv Detail & Related papers (2020-09-15T16:33:05Z) - Deep Active Learning by Model Interpretability [7.3461534048332275]
In this paper, we introduce the linearly separable regions of samples to the problem of active learning.
We propose a novel Deep Active learning approach by Model Interpretability (DAMI)
To keep the maximal representativeness of the entire unlabeled data, DAMI tries to select and label samples on different linearly separable regions.
arXiv Detail & Related papers (2020-07-23T16:06:27Z) - Stance Detection Benchmark: How Robust Is Your Stance Detection? [65.91772010586605]
Stance Detection (StD) aims to detect an author's stance towards a certain topic or claim.
We introduce a StD benchmark that learns from ten StD datasets of various domains in a multi-dataset learning setting.
Within this benchmark setup, we are able to present new state-of-the-art results on five of the datasets.
arXiv Detail & Related papers (2020-01-06T13:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.