Probabilistic Diagnostic Tests for Degradation Problems in Supervised
Learning
- URL: http://arxiv.org/abs/2004.02988v2
- Date: Wed, 15 Apr 2020 19:12:24 GMT
- Title: Probabilistic Diagnostic Tests for Degradation Problems in Supervised
Learning
- Authors: Gustavo A. Valencia-Zapata, Carolina Gonzalez-Canas, Michael G.
Zentner, Okan Ersoy, and Gerhard Klimeck
- Abstract summary: Problems such as class imbalance, overlapping, small-disjuncts, noisy labels, and sparseness limit accuracy in classification algorithms.
Probability diagnostic model based on identifying signs and symptoms of each problem is presented.
Behavior and performance of several supervised algorithms are studied when training sets have such problems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Several studies point out different causes of performance degradation in
supervised machine learning. Problems such as class imbalance, overlapping,
small-disjuncts, noisy labels, and sparseness limit accuracy in classification
algorithms. Even though a number of approaches either in the form of a
methodology or an algorithm try to minimize performance degradation, they have
been isolated efforts with limited scope. Most of these approaches focus on
remediation of one among many problems, with experimental results coming from
few datasets and classification algorithms, insufficient measures of prediction
power, and lack of statistical validation for testing the real benefit of the
proposed approach. This paper consists of two main parts: In the first part, a
novel probabilistic diagnostic model based on identifying signs and symptoms of
each problem is presented. Thereby, early and correct diagnosis of these
problems is to be achieved in order to select not only the most convenient
remediation treatment but also unbiased performance metrics. Secondly, the
behavior and performance of several supervised algorithms are studied when
training sets have such problems. Therefore, prediction of success for
treatments can be estimated across classifiers.
Related papers
- Predictor-Rejector Multi-Class Abstention: Theoretical Analysis and Algorithms [30.389055604165222]
We study the key framework of learning with abstention in the multi-class classification setting.
In this setting, the learner can choose to abstain from making a prediction with some pre-defined cost.
We introduce several new families of surrogate losses for which we prove strong non-asymptotic and hypothesis set-specific consistency guarantees.
arXiv Detail & Related papers (2023-10-23T10:16:27Z) - A Unified Generalization Analysis of Re-Weighting and Logit-Adjustment
for Imbalanced Learning [129.63326990812234]
We propose a technique named data-dependent contraction to capture how modified losses handle different classes.
On top of this technique, a fine-grained generalization bound is established for imbalanced learning, which helps reveal the mystery of re-weighting and logit-adjustment.
arXiv Detail & Related papers (2023-10-07T09:15:08Z) - Predictive Coding beyond Correlations [59.47245250412873]
We show how one of such algorithms, called predictive coding, is able to perform causal inference tasks.
First, we show how a simple change in the inference process of predictive coding enables to compute interventions without the need to mutilate or redefine a causal graph.
arXiv Detail & Related papers (2023-06-27T13:57:16Z) - Adaptive Learning for the Resource-Constrained Classification Problem [14.19197444541245]
Resource-constrained classification tasks are common in real-world applications such as allocating tests for disease diagnosis.
We design an adaptive learning approach that considers resource constraints and learning jointly by iteratively fine-tuning misclassification costs.
We envision the adaptive learning approach as an important addition to the repertoire of techniques for handling resource-constrained classification problems.
arXiv Detail & Related papers (2022-07-19T11:00:33Z) - Towards Diverse Evaluation of Class Incremental Learning: A Representation Learning Perspective [67.45111837188685]
Class incremental learning (CIL) algorithms aim to continually learn new object classes from incrementally arriving data.
We experimentally analyze neural network models trained by CIL algorithms using various evaluation protocols in representation learning.
arXiv Detail & Related papers (2022-06-16T11:44:11Z) - Learning to Rank Anomalies: Scalar Performance Criteria and Maximization
of Two-Sample Rank Statistics [0.0]
We propose a data-driven scoring function defined on the feature space which reflects the degree of abnormality of the observations.
This scoring function is learnt through a well-designed binary classification problem.
We illustrate our methodology with preliminary encouraging numerical experiments.
arXiv Detail & Related papers (2021-09-20T14:45:56Z) - Few-shot Action Recognition with Prototype-centered Attentive Learning [88.10852114988829]
Prototype-centered Attentive Learning (PAL) model composed of two novel components.
First, a prototype-centered contrastive learning loss is introduced to complement the conventional query-centered learning objective.
Second, PAL integrates a attentive hybrid learning mechanism that can minimize the negative impacts of outliers.
arXiv Detail & Related papers (2021-01-20T11:48:12Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Cross-validation Confidence Intervals for Test Error [83.67415139421448]
This work develops central limit theorems for crossvalidation and consistent estimators of its variance under weak stability conditions on the learning algorithm.
Results are the first of their kind for the popular choice of leave-one-out cross-validation.
arXiv Detail & Related papers (2020-07-24T17:40:06Z) - Better Multi-class Probability Estimates for Small Data Sets [0.0]
We show that Data Generation and Grouping algorithm can be used to solve multi-class problems.
Our experiments show that calibration error can be decreased using the proposed approach and the additional computational cost is acceptable.
arXiv Detail & Related papers (2020-01-30T10:21:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.