Ensemble Learning using Error Correcting Output Codes: New
Classification Error Bounds
- URL: http://arxiv.org/abs/2109.08967v1
- Date: Sat, 18 Sep 2021 16:47:57 GMT
- Title: Ensemble Learning using Error Correcting Output Codes: New
Classification Error Bounds
- Authors: Hieu D. Nguyen, Mohammed Sarosh Khan, Nicholas Kaegi, Shen-Shyang Ho,
Jonathan Moore, Logan Borys, Lucas Lavalva
- Abstract summary: We present new bounds on classification error rates for the error-correcting output code (ECOC) approach in machine learning.
These bounds have exponential decay complexity with respect to codeword length and theoretically validate the effectiveness of the ECOC approach.
- Score: 2.0242396022517752
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: New bounds on classification error rates for the error-correcting output code
(ECOC) approach in machine learning are presented. These bounds have
exponential decay complexity with respect to codeword length and theoretically
validate the effectiveness of the ECOC approach. Bounds are derived for two
different models: the first under the assumption that all base classifiers are
independent and the second under the assumption that all base classifiers are
mutually correlated up to first-order. Moreover, we perform ECOC classification
on six datasets and compare their error rates with our bounds to experimentally
validate our work and show the effect of correlation on classification
accuracy.
Related papers
- Classification Error Bound for Low Bayes Error Conditions in Machine Learning [50.25063912757367]
We study the relationship between the error mismatch and the Kullback-Leibler divergence in machine learning.
Motivated by recent observations of low model-based classification errors in many machine learning tasks, we propose a linear approximation of the classification error bound for low Bayes error conditions.
arXiv Detail & Related papers (2025-01-27T11:57:21Z) - Risk-based Calibration for Generative Classifiers [4.792851066169872]
We propose a learning procedure called risk-based calibration (RC)
RC iteratively refines the generative classifier by adjusting its joint probability distribution according to the 0-1 loss in training samples.
RC significantly outperforms closed-form learning procedures in terms of both training error and generalization error.
arXiv Detail & Related papers (2024-09-05T14:06:56Z) - Adapting Conformal Prediction to Distribution Shifts Without Labels [16.478151550456804]
Conformal prediction (CP) enables machine learning models to output prediction sets with guaranteed coverage rate.
Our goal is to improve the quality of CP-generated prediction sets using only unlabeled data from the test domain.
This is achieved by two new methods called ECP and EACP, that adjust the score function in CP according to the base model's uncertainty on the unlabeled test data.
arXiv Detail & Related papers (2024-06-03T15:16:02Z) - Data-driven Error Estimation: Upper Bounding Multiple Errors without Class Complexity as Input [24.549751508566565]
We frame this as an "error estimation problem"
The goal is to determine a high-probability upper bound on the maximum error for a class of estimates.
We propose an entirely data-driven approach that derives such bounds for both finite and infinite class settings.
arXiv Detail & Related papers (2024-05-07T19:38:26Z) - JointMatch: A Unified Approach for Diverse and Collaborative
Pseudo-Labeling to Semi-Supervised Text Classification [65.268245109828]
Semi-supervised text classification (SSTC) has gained increasing attention due to its ability to leverage unlabeled data.
Existing approaches based on pseudo-labeling suffer from the issues of pseudo-label bias and error accumulation.
We propose JointMatch, a holistic approach for SSTC that addresses these challenges by unifying ideas from recent semi-supervised learning.
arXiv Detail & Related papers (2023-10-23T05:43:35Z) - Fair Canonical Correlation Analysis [14.206538828733507]
Canonical Correlation Analysis (CCA) is a widely used technique for examining the relationship between two sets of variables.
We present a framework that alleviates unfairness by minimizing the correlation disparity error associated with protected attributes.
arXiv Detail & Related papers (2023-09-27T17:34:13Z) - Deep Negative Correlation Classification [82.45045814842595]
Existing deep ensemble methods naively train many different models and then aggregate their predictions.
We propose deep negative correlation classification (DNCC)
DNCC yields a deep classification ensemble where the individual estimator is both accurate and negatively correlated.
arXiv Detail & Related papers (2022-12-14T07:35:20Z) - Complementary Labels Learning with Augmented Classes [22.460256396941528]
Complementary Labels Learning (CLL) arises in many real-world tasks such as private questions classification and online learning.
We propose a novel problem setting called Complementary Labels Learning with Augmented Classes (CLLAC)
By using unlabeled data, we propose an unbiased estimator of classification risk for CLLAC, which is guaranteed to be provably consistent.
arXiv Detail & Related papers (2022-11-19T13:55:27Z) - Mutual Information Learned Classifiers: an Information-theoretic
Viewpoint of Training Deep Learning Classification Systems [9.660129425150926]
We show that the existing cross entropy loss minimization problem essentially learns the label conditional entropy of the underlying data distribution.
We propose a mutual information learning framework where we train deep neural network classifiers via learning the mutual information between the label and the input.
arXiv Detail & Related papers (2022-09-21T01:06:30Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Information-Theoretic Generalization Bounds for Iterative
Semi-Supervised Learning [81.1071978288003]
In particular, we seek to understand the behaviour of the em generalization error of iterative SSL algorithms using information-theoretic principles.
Our theoretical results suggest that when the class conditional variances are not too large, the upper bound on the generalization error decreases monotonically with the number of iterations, but quickly saturates.
arXiv Detail & Related papers (2021-10-03T05:38:49Z) - How to Control the Error Rates of Binary Classifiers [0.0]
We show how to turn binary classification into statistical tests, calculate the classification p-values, and use them to limit the target error rate.
In particular, we show how to turn binary classifiers into statistical tests, calculate the classification p-values, and use them to limit the target error rate.
arXiv Detail & Related papers (2020-10-21T14:43:14Z) - Learning by Minimizing the Sum of Ranked Range [58.24935359348289]
We introduce the sum of ranked range (SoRR) as a general approach to form learning objectives.
A ranked range is a consecutive sequence of sorted values of a set of real numbers.
We explore two applications in machine learning of the minimization of the SoRR framework, namely the AoRR aggregate loss for binary classification and the TKML individual loss for multi-label/multi-class classification.
arXiv Detail & Related papers (2020-10-05T01:58:32Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.