The Optimal Input-Independent Baseline for Binary Classification: The
Dutch Draw
- URL: http://arxiv.org/abs/2301.03318v1
- Date: Mon, 9 Jan 2023 13:11:59 GMT
- Title: The Optimal Input-Independent Baseline for Binary Classification: The
Dutch Draw
- Authors: Joris Pries, Etienne van de Bijl, Jan Klein, Sandjai Bhulai, Rob van
der Mei
- Abstract summary: The goal of this paper is to examine all baseline methods that are independent of feature values.
By identifying which baseline models are optimal, a crucial selection decision in the evaluation process is simplified.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Before any binary classification model is taken into practice, it is
important to validate its performance on a proper test set. Without a frame of
reference given by a baseline method, it is impossible to determine if a score
is `good' or `bad'. The goal of this paper is to examine all baseline methods
that are independent of feature values and determine which model is the `best'
and why. By identifying which baseline models are optimal, a crucial selection
decision in the evaluation process is simplified. We prove that the recently
proposed Dutch Draw baseline is the best input-independent classifier
(independent of feature values) for all positional-invariant measures
(independent of sequence order) assuming that the samples are randomly
shuffled. This means that the Dutch Draw baseline is the optimal baseline under
these intuitive requirements and should therefore be used in practice.
Related papers
- BAPE: Learning an Explicit Bayes Classifier for Long-tailed Visual Recognition [78.70453964041718]
Current deep learning algorithms usually solve for the optimal classifier by emphimplicitly estimating the posterior probabilities.<n>This simple methodology has been proven effective for meticulously balanced academic benchmark datasets.<n>However, it is not applicable to the long-tailed data distributions in the real world.<n>This paper presents a novel approach (BAPE) that provides a more precise theoretical estimation of the data distributions.
arXiv Detail & Related papers (2025-06-29T15:12:50Z) - Achieving More with Less: A Tensor-Optimization-Powered Ensemble Method [53.170053108447455]
Ensemble learning is a method that leverages weak learners to produce a strong learner.
We design a smooth and convex objective function that leverages the concept of margin, making the strong learner more discriminative.
We then compare our algorithm with random forests of ten times the size and other classical methods across numerous datasets.
arXiv Detail & Related papers (2024-08-06T03:42:38Z) - Optimal Baseline Corrections for Off-Policy Contextual Bandits [61.740094604552475]
We aim to learn decision policies that optimize an unbiased offline estimate of an online reward metric.
We propose a single framework built on their equivalence in learning scenarios.
Our framework enables us to characterize the variance-optimal unbiased estimator and provide a closed-form solution for it.
arXiv Detail & Related papers (2024-05-09T12:52:22Z) - Leveraging Uncertainty Estimates To Improve Classifier Performance [4.4951754159063295]
Binary classification involves predicting the label of an instance based on whether the model score for the positive class exceeds a threshold chosen based on the application requirements.
However, model scores are often not aligned with the true positivity rate.
This is especially true when the training involves a differential sampling across classes or there is distributional drift between train and test settings.
arXiv Detail & Related papers (2023-11-20T12:40:25Z) - The Dutch Draw: Constructing a Universal Baseline for Binary Prediction
Models [2.8816551600116527]
A proper baseline is needed to evaluate the goodness' of a performance score.
This paper presents a universal baseline method for all binary classification models, named the Dutch Draw (DD)
arXiv Detail & Related papers (2022-03-24T14:20:27Z) - Scalable Optimal Classifiers for Adversarial Settings under Uncertainty [10.90668635921398]
We consider the problem of finding optimal classifiers in an adversarial setting where the class-1 data is generated by an attacker whose objective is not known to the defender.
We show that this low-dimensional characterization enables to develop a training method to compute provably approximately optimal classifiers in a scalable manner.
arXiv Detail & Related papers (2021-06-28T13:33:53Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - A Baseline for Shapley Values in MLPs: from Missingness to Neutrality [3.5939555573102853]
Deep neural networks have gained momentum based on their accuracy, but their interpretability is often criticised.
In this paper, we present a method for choosing a baseline according to a neutrality value.
We empirically validate our choice of baseline in the context of binary classification tasks.
arXiv Detail & Related papers (2020-06-08T19:29:36Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z) - Learning to Select Base Classes for Few-shot Classification [96.92372639495551]
We use the Similarity Ratio as an indicator for the generalization performance of a few-shot model.
We then formulate the base class selection problem as a submodular optimization problem over Similarity Ratio.
arXiv Detail & Related papers (2020-04-01T09:55:18Z) - Progressive Identification of True Labels for Partial-Label Learning [112.94467491335611]
Partial-label learning (PLL) is a typical weakly supervised learning problem, where each training instance is equipped with a set of candidate labels among which only one is the true label.
Most existing methods elaborately designed as constrained optimizations that must be solved in specific manners, making their computational complexity a bottleneck for scaling up to big data.
This paper proposes a novel framework of classifier with flexibility on the model and optimization algorithm.
arXiv Detail & Related papers (2020-02-19T08:35:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.