On robust risk-based active-learning algorithms for enhanced decision
support
- URL: http://arxiv.org/abs/2201.02555v1
- Date: Fri, 7 Jan 2022 17:25:41 GMT
- Title: On robust risk-based active-learning algorithms for enhanced decision
support
- Authors: Aidan J. Hughes, Lawrence A. Bull, Paul Gardner, Nikolaos Dervilis,
Keith Worden
- Abstract summary: Classification models are a fundamental component of physical-asset management technologies such as structural health monitoring (SHM) systems and digital twins.
The paper proposes two novel approaches to counteract the effects of sampling bias: textitsemi-supervised learning, and textitdiscriminative classification models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Classification models are a fundamental component of physical-asset
management technologies such as structural health monitoring (SHM) systems and
digital twins. Previous work introduced \textit{risk-based active learning}, an
online approach for the development of statistical classifiers that takes into
account the decision-support context in which they are applied. Decision-making
is considered by preferentially querying data labels according to
\textit{expected value of perfect information} (EVPI). Although several
benefits are gained by adopting a risk-based active learning approach,
including improved decision-making performance, the algorithms suffer from
issues relating to sampling bias as a result of the guided querying process.
This sampling bias ultimately manifests as a decline in decision-making
performance during the later stages of active learning, which in turn
corresponds to lost resource/utility.
The current paper proposes two novel approaches to counteract the effects of
sampling bias: \textit{semi-supervised learning}, and \textit{discriminative
classification models}. These approaches are first visualised using a synthetic
dataset, then subsequently applied to an experimental case study, specifically,
the Z24 Bridge dataset. The semi-supervised learning approach is shown to have
variable performance; with robustness to sampling bias dependent on the
suitability of the generative distributions selected for the model with respect
to each dataset. In contrast, the discriminative classifiers are shown to have
excellent robustness to the effects of sampling bias. Moreover, it was found
that the number of inspections made during a monitoring campaign, and therefore
resource expenditure, could be reduced with the careful selection of the
statistical classifiers used within a decision-supporting monitoring system.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - DRoP: Distributionally Robust Pruning [11.930434318557156]
We conduct the first systematic study of the impact of data pruning on classification bias of trained models.
We propose DRoP, a distributionally robust approach to pruning and empirically demonstrate its performance on standard computer vision benchmarks.
arXiv Detail & Related papers (2024-04-08T14:55:35Z) - Learning with Complementary Labels Revisited: The Selected-Completely-at-Random Setting Is More Practical [66.57396042747706]
Complementary-label learning is a weakly supervised learning problem.
We propose a consistent approach that does not rely on the uniform distribution assumption.
We find that complementary-label learning can be expressed as a set of negative-unlabeled binary classification problems.
arXiv Detail & Related papers (2023-11-27T02:59:17Z) - Optimal Sample Selection Through Uncertainty Estimation and Its
Application in Deep Learning [22.410220040736235]
We present a theoretically optimal solution for addressing both coreset selection and active learning.
Our proposed method, COPS, is designed to minimize the expected loss of a model trained on subsampled data.
arXiv Detail & Related papers (2023-09-05T14:06:33Z) - Delving into Identify-Emphasize Paradigm for Combating Unknown Bias [52.76758938921129]
We propose an effective bias-conflicting scoring method (ECS) to boost the identification accuracy.
We also propose gradient alignment (GA) to balance the contributions of the mined bias-aligned and bias-conflicting samples.
Experiments are conducted on multiple datasets in various settings, demonstrating that the proposed solution can mitigate the impact of unknown biases.
arXiv Detail & Related papers (2023-02-22T14:50:24Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - Mitigating sampling bias in risk-based active learning via an EM
algorithm [0.0]
Risk-based active learning is an approach to developing statistical classifiers for online decision-support.
Data-label querying is guided according to the expected value of perfect information for incipient data points.
Semi-supervised approach counteracts sampling bias by incorporating pseudo-labels for unlabelled data via an EM algorithm.
arXiv Detail & Related papers (2022-06-25T08:48:25Z) - Improving decision-making via risk-based active learning: Probabilistic
discriminative classifiers [0.0]
descriptive labels for measured data corresponding to health-states of monitored systems are often unavailable.
One approach to dealing with this problem is risk-based active learning.
The current paper demonstrates several advantages of using an alternative type of classifier -- discriminative models.
arXiv Detail & Related papers (2022-06-23T10:51:42Z) - Learning Bias-Invariant Representation by Cross-Sample Mutual
Information Minimization [77.8735802150511]
We propose a cross-sample adversarial debiasing (CSAD) method to remove the bias information misused by the target task.
The correlation measurement plays a critical role in adversarial debiasing and is conducted by a cross-sample neural mutual information estimator.
We conduct thorough experiments on publicly available datasets to validate the advantages of the proposed method over state-of-the-art approaches.
arXiv Detail & Related papers (2021-08-11T21:17:02Z) - Deep Active Learning for Solvability Prediction in Power Systems [7.634675607430369]
We propose a deep active learning framework for power system solvability prediction.
Compared with the passive learning methods where the training is performed after all instances are labeled, the active learning selects most informative instances to be label.
arXiv Detail & Related papers (2020-07-27T00:13:09Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.