On the Effects of Randomness on Stability of Learning with Limited
Labelled Data: A Systematic Literature Review
- URL: http://arxiv.org/abs/2312.01082v1
- Date: Sat, 2 Dec 2023 09:20:10 GMT
- Title: On the Effects of Randomness on Stability of Learning with Limited
Labelled Data: A Systematic Literature Review
- Authors: Branislav Pecher, Ivan Srba, Maria Bielikova
- Abstract summary: This paper provides a comprehensive overview of 134 papers addressing the effects of randomness on the stability of learning with limited labelled data.
We identify and discuss seven challenges and open problems together with possible directions to facilitate further research.
- Score: 5.630038762653309
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning with limited labelled data, such as few-shot learning, meta-learning
or transfer learning, aims to effectively train a model using only small amount
of labelled samples. However, these approaches were observed to be excessively
sensitive to the effects of uncontrolled randomness caused by non-determinism
in the training process. The randomness negatively affects the stability of the
models, leading to large variance in results across training runs. When such
instability is disregarded, it can unintentionally, but unfortunately also
intentionally, create an imaginary perception of research progress. Recently,
this area started to attract a research attention and the number of relevant
studies is continuously growing. In this survey, we provide a comprehensive
overview of 134 papers addressing the effects of randomness on the stability of
learning with limited labelled data. We distinguish between four main tasks
addressed in the papers (investigate/evaluate; determine; mitigate;
benchmark/compare/report randomness effects), providing findings for each one.
Furthermore, we identify and discuss seven challenges and open problems
together with possible directions to facilitate further research. The ultimate
goal of this survey is to emphasise the importance of this growing research
area, which so far has not received appropriate level of attention.
Related papers
- On Sensitivity of Learning with Limited Labelled Data to the Effects of
Randomness: Impact of Interactions and Systematic Choices [5.630038762653309]
We propose a method to investigate the effects of randomness factors while taking the interactions into consideration.
To measure the true effects of an individual randomness factor, our method mitigates the effects of other factors and observes how the performance varies across multiple runs.
Applying our method to multiple randomness factors across in-context learning and fine-tuning approaches on 7 representative text classification tasks and meta-learning on 3 tasks, we show that: 1) disregarding interactions between randomness factors in existing works caused inconsistent findings due to incorrect attribution of the effects of randomness factors, such as disproving the consistent sensitivity of in-context learning to sample order even
arXiv Detail & Related papers (2024-02-20T08:38:19Z) - Assumption violations in causal discovery and the robustness of score
matching [40.269084132312706]
This paper extensively benchmarks the empirical performance of recent causal discovery methods on observational i.i.d. data.
We show that score matching-based methods demonstrate surprising performance in the false positive and false negative rate of the inferred graph.
We hope this paper will set a new standard for the evaluation of causal discovery methods.
arXiv Detail & Related papers (2023-10-20T09:56:07Z) - Seeing is not Believing: Robust Reinforcement Learning against Spurious
Correlation [57.351098530477124]
We consider one critical type of robustness against spurious correlation, where different portions of the state do not have correlations induced by unobserved confounders.
A model that learns such useless or even harmful correlation could catastrophically fail when the confounder in the test case deviates from the training one.
Existing robust algorithms that assume simple and unstructured uncertainty sets are therefore inadequate to address this challenge.
arXiv Detail & Related papers (2023-07-15T23:53:37Z) - Leveraging Unlabelled Data in Multiple-Instance Learning Problems for
Improved Detection of Parkinsonian Tremor in Free-Living Conditions [80.88681952022479]
We introduce a new method for combining semi-supervised with multiple-instance learning.
We show that by leveraging the unlabelled data of 454 subjects we can achieve large performance gains in per-subject tremor detection.
arXiv Detail & Related papers (2023-04-29T12:25:10Z) - ALLSH: Active Learning Guided by Local Sensitivity and Hardness [98.61023158378407]
We propose to retrieve unlabeled samples with a local sensitivity and hardness-aware acquisition function.
Our method achieves consistent gains over the commonly used active learning strategies in various classification tasks.
arXiv Detail & Related papers (2022-05-10T15:39:11Z) - Partial Identification of Dose Responses with Hidden Confounders [25.468473751289036]
Inferring causal effects of continuous-valued treatments from observational data is a crucial task.
We present novel methodology to bound both average and conditional average continuous-valued treatment-effect estimates.
We apply our method to a real-world observational case study to demonstrate the value of identifying dose-dependent causal effects.
arXiv Detail & Related papers (2022-04-24T07:02:21Z) - SurvITE: Learning Heterogeneous Treatment Effects from Time-to-Event
Data [83.50281440043241]
We study the problem of inferring heterogeneous treatment effects from time-to-event data.
We propose a novel deep learning method for treatment-specific hazard estimation based on balancing representations.
arXiv Detail & Related papers (2021-10-26T20:13:17Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Unsupervised Embedding Learning from Uncertainty Momentum Modeling [37.674449317054716]
We propose a novel solution to explicitly model and explore the uncertainty of the given unlabeled learning samples.
We leverage such uncertainty modeling momentum to the learning which is helpful to tackle the outliers.
arXiv Detail & Related papers (2021-07-19T14:06:19Z) - Bayesian Active Learning for Wearable Stress and Affect Detection [0.7106986689736827]
Stress detection using on-device deep learning algorithms has been on the rise owing to advancements in pervasive computing.
In this paper, we propose a framework with capabilities to represent model uncertainties through approximations in Bayesian Neural Networks.
Our proposed framework achieves a considerable efficiency boost during inference, with a substantially low number of acquired pool points.
arXiv Detail & Related papers (2020-12-04T16:19:37Z) - Hidden Cost of Randomized Smoothing [72.93630656906599]
In this paper, we point out the side effects of current randomized smoothing.
Specifically, we articulate and prove two major points: 1) the decision boundaries of smoothed classifiers will shrink, resulting in disparity in class-wise accuracy; 2) applying noise augmentation in the training process does not necessarily resolve the shrinking issue due to the inconsistent learning objectives.
arXiv Detail & Related papers (2020-03-02T23:37:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.