Performance-Agnostic Fusion of Probabilistic Classifier Outputs
- URL: http://arxiv.org/abs/2009.00565v1
- Date: Tue, 1 Sep 2020 16:53:29 GMT
- Title: Performance-Agnostic Fusion of Probabilistic Classifier Outputs
- Authors: Jordan F. Masakuna, Simukai W. Utete, Steve Kroon
- Abstract summary: We propose a method for combining probabilistic outputs of classifiers to make a single consensus class prediction.
Our proposed method works well in situations where accuracy is the performance metric.
It does not output calibrated probabilities, so it is not suitable in situations where such probabilities are required for further processing.
- Score: 2.4206828137867107
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a method for combining probabilistic outputs of classifiers to
make a single consensus class prediction when no further information about the
individual classifiers is available, beyond that they have been trained for the
same task. The lack of relevant prior information rules out typical
applications of Bayesian or Dempster-Shafer methods, and the default approach
here would be methods based on the principle of indifference, such as the sum
or product rule, which essentially weight all classifiers equally. In contrast,
our approach considers the diversity between the outputs of the various
classifiers, iteratively updating predictions based on their correspondence
with other predictions until the predictions converge to a consensus decision.
The intuition behind this approach is that classifiers trained for the same
task should typically exhibit regularities in their outputs on a new task; the
predictions of classifiers which differ significantly from those of others are
thus given less credence using our approach. The approach implicitly assumes a
symmetric loss function, in that the relative cost of various prediction errors
are not taken into account. Performance of the model is demonstrated on
different benchmark datasets. Our proposed method works well in situations
where accuracy is the performance metric; however, it does not output
calibrated probabilities, so it is not suitable in situations where such
probabilities are required for further processing.
Related papers
- Trustworthy Classification through Rank-Based Conformal Prediction Sets [9.559062601251464]
We propose a novel conformal prediction method that employs a rank-based score function suitable for classification models.
Our approach constructs prediction sets that achieve the desired coverage rate while managing their size.
Our contributions include a novel conformal prediction method, theoretical analysis, and empirical evaluation.
arXiv Detail & Related papers (2024-07-05T10:43:41Z) - Invariant Probabilistic Prediction [45.90606906307022]
We show that arbitrary distribution shifts do not, in general, admit invariant and robust probabilistic predictions.
We propose a method to yield invariant probabilistic predictions, called IPP, and study the consistency of the underlying parameters.
arXiv Detail & Related papers (2023-09-18T18:50:24Z) - Quantification of Predictive Uncertainty via Inference-Time Sampling [57.749601811982096]
We propose a post-hoc sampling strategy for estimating predictive uncertainty accounting for data ambiguity.
The method can generate different plausible outputs for a given input and does not assume parametric forms of predictive distributions.
arXiv Detail & Related papers (2023-08-03T12:43:21Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - From Classification Accuracy to Proper Scoring Rules: Elicitability of
Probabilistic Top List Predictions [0.0]
I propose a novel type of prediction in classification, which bridges the gap between single-class predictions and predictive distributions.
The proposed evaluation metrics are based on symmetric proper scoring rules and admit comparison of various types of predictions.
arXiv Detail & Related papers (2023-01-27T15:55:01Z) - Predicting Out-of-Domain Generalization with Neighborhood Invariance [59.05399533508682]
We propose a measure of a classifier's output invariance in a local transformation neighborhood.
Our measure is simple to calculate, does not depend on the test point's true label, and can be applied even in out-of-domain (OOD) settings.
In experiments on benchmarks in image classification, sentiment analysis, and natural language inference, we demonstrate a strong and robust correlation between our measure and actual OOD generalization.
arXiv Detail & Related papers (2022-07-05T14:55:16Z) - Self-Certifying Classification by Linearized Deep Assignment [65.0100925582087]
We propose a novel class of deep predictors for classifying metric data on graphs within PAC-Bayes risk certification paradigm.
Building on the recent PAC-Bayes literature and data-dependent priors, this approach enables learning posterior distributions on the hypothesis space.
arXiv Detail & Related papers (2022-01-26T19:59:14Z) - When in Doubt: Improving Classification Performance with Alternating
Normalization [57.39356691967766]
We introduce Classification with Alternating Normalization (CAN), a non-parametric post-processing step for classification.
CAN improves classification accuracy for challenging examples by re-adjusting their predicted class probability distribution.
We empirically demonstrate its effectiveness across a diverse set of classification tasks.
arXiv Detail & Related papers (2021-09-28T02:55:42Z) - Optimal strategies for reject option classifiers [0.0]
In classification with a reject option, the classifier is allowed in uncertain cases to abstain from prediction.
We coin a symmetric definition, the bounded-coverage model, which seeks for a classifier with minimal selective risk and guaranteed coverage.
We propose two algorithms to learn the proper uncertainty score from examples for an arbitrary black-box classifier.
arXiv Detail & Related papers (2021-01-29T11:09:32Z) - Certified Robustness to Label-Flipping Attacks via Randomized Smoothing [105.91827623768724]
Machine learning algorithms are susceptible to data poisoning attacks.
We present a unifying view of randomized smoothing over arbitrary functions.
We propose a new strategy for building classifiers that are pointwise-certifiably robust to general data poisoning attacks.
arXiv Detail & Related papers (2020-02-07T21:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.