Minimax Classification with 0-1 Loss and Performance Guarantees
- URL: http://arxiv.org/abs/2010.07964v1
- Date: Thu, 15 Oct 2020 18:11:28 GMT
- Title: Minimax Classification with 0-1 Loss and Performance Guarantees
- Authors: Santiago Mazuelas and Andrea Zanoni and Aritz Perez
- Abstract summary: Supervised classification techniques use training samples to find classification rules with small expected 0-1 loss.
Conventional methods achieve efficient learning and out-of-sample generalization by minimizing surrogate losses over specific families of rules.
This paper presents minimax risk classifiers (MRCs) that do not rely on a choice of surrogate loss and family of rules.
- Score: 4.812718493682455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised classification techniques use training samples to find
classification rules with small expected 0-1 loss. Conventional methods achieve
efficient learning and out-of-sample generalization by minimizing surrogate
losses over specific families of rules. This paper presents minimax risk
classifiers (MRCs) that do not rely on a choice of surrogate loss and family of
rules. MRCs achieve efficient learning and out-of-sample generalization by
minimizing worst-case expected 0-1 loss w.r.t. uncertainty sets that are
defined by linear constraints and include the true underlying distribution. In
addition, MRCs' learning stage provides performance guarantees as lower and
upper tight bounds for expected 0-1 loss. We also present MRCs' finite-sample
generalization bounds in terms of training size and smallest minimax risk, and
show their competitive classification performance w.r.t. state-of-the-art
techniques using benchmark datasets.
Related papers
- Learning Towards the Largest Margins [83.7763875464011]
Loss function should promote the largest possible margins for both classes and samples.
Not only does this principled framework offer new perspectives to understand and interpret existing margin-based losses, but it can guide the design of new tools.
arXiv Detail & Related papers (2022-06-23T10:03:03Z) - Learning Imbalanced Datasets with Maximum Margin Loss [21.305656747991026]
A learning algorithm referred to as Maximum Margin (MM) is proposed for considering the class-imbalance data learning issue.
We design a new Maximum Margin (MM) loss function, motivated by minimizing a margin-based generalization bound through the shifting decision bound.
arXiv Detail & Related papers (2022-06-11T00:21:41Z) - Minimax risk classifiers with 0-1 loss [7.650319416775203]
This paper presents minimax risk classifiers (MRCs) that minize the worst-case 0-1 loss with respect to uncertainty sets of distributions.
We show that MRCs can provide tight performance guarantees at learning and are strongly universally consistent using feature mappings given by characteristic kernels.
The paper also proposes efficient optimization techniques for MRC learning and shows that the methods presented can provide accurate classification together with tight performance guarantees in practice.
arXiv Detail & Related papers (2022-01-17T16:00:07Z) - Shift Happens: Adjusting Classifiers [2.8682942808330703]
Minimizing expected loss measured by a proper scoring rule, such as Brier score or log-loss (cross-entropy), is a common objective while training a probabilistic classifier.
We propose methods that transform all predictions to (re)equalize the average prediction and the class distribution.
We demonstrate experimentally that, when in practice the class distribution is known only approximately, there is often still a reduction in loss depending on the amount of shift and the precision to which the class distribution is known.
arXiv Detail & Related papers (2021-11-03T21:27:27Z) - Learning with Multiclass AUC: Theory and Algorithms [141.63211412386283]
Area under the ROC curve (AUC) is a well-known ranking metric for problems such as imbalanced learning and recommender systems.
In this paper, we start an early trial to consider the problem of learning multiclass scoring functions via optimizing multiclass AUC metrics.
arXiv Detail & Related papers (2021-07-28T05:18:10Z) - Constrained Classification and Policy Learning [0.0]
We study consistency of surrogate loss procedures under a constrained set of classifiers.
We show that hinge losses are the only surrogate losses that preserve consistency in second-best scenarios.
arXiv Detail & Related papers (2021-06-24T10:43:00Z) - Risk Minimization from Adaptively Collected Data: Guarantees for
Supervised and Policy Learning [57.88785630755165]
Empirical risk minimization (ERM) is the workhorse of machine learning, but its model-agnostic guarantees can fail when we use adaptively collected data.
We study a generic importance sampling weighted ERM algorithm for using adaptively collected data to minimize the average of a loss function over a hypothesis class.
For policy learning, we provide rate-optimal regret guarantees that close an open gap in the existing literature whenever exploration decays to zero.
arXiv Detail & Related papers (2021-06-03T09:50:13Z) - Learning, compression, and leakage: Minimising classification error via
meta-universal compression principles [87.054014983402]
A promising group of compression techniques for learning scenarios is normalised maximum likelihood (NML) coding.
Here we consider a NML-based decision strategy for supervised classification problems, and show that it attains PAC learning when applied to a wide variety of models.
We show that the misclassification rate of our method is upper bounded by the maximal leakage, a recently proposed metric to quantify the potential of data leakage in privacy-sensitive scenarios.
arXiv Detail & Related papers (2020-10-14T20:03:58Z) - Learning by Minimizing the Sum of Ranked Range [58.24935359348289]
We introduce the sum of ranked range (SoRR) as a general approach to form learning objectives.
A ranked range is a consecutive sequence of sorted values of a set of real numbers.
We explore two applications in machine learning of the minimization of the SoRR framework, namely the AoRR aggregate loss for binary classification and the TKML individual loss for multi-label/multi-class classification.
arXiv Detail & Related papers (2020-10-05T01:58:32Z) - Calibrated Surrogate Losses for Adversarially Robust Classification [92.37268323142307]
We show that no convex surrogate loss is respect with respect to adversarial 0-1 loss when restricted to linear models.
We also show that if the underlying distribution satisfies the Massart's noise condition, convex losses can also be calibrated in the adversarial setting.
arXiv Detail & Related papers (2020-05-28T02:40:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.