FPR Estimation for Fraud Detection in the Presence of Class-Conditional
Label Noise
- URL: http://arxiv.org/abs/2308.02695v1
- Date: Fri, 4 Aug 2023 20:14:34 GMT
- Title: FPR Estimation for Fraud Detection in the Presence of Class-Conditional
Label Noise
- Authors: Justin Tittelfitz
- Abstract summary: We consider the problem of estimating the false-/ true-positive-rate (FPR/TPR) for a binary classification model when there are incorrect labels (label noise) in the validation set.
Our motivating application is fraud prevention where accurate estimates of FPR are critical to preserving the experience for good customers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of estimating the false-/ true-positive-rate
(FPR/TPR) for a binary classification model when there are incorrect labels
(label noise) in the validation set. Our motivating application is fraud
prevention where accurate estimates of FPR are critical to preserving the
experience for good customers, and where label noise is highly asymmetric.
Existing methods seek to minimize the total error in the cleaning process - to
avoid cleaning examples that are not noise, and to ensure cleaning of examples
that are. This is an important measure of accuracy but insufficient to
guarantee good estimates of the true FPR or TPR for a model, and we show that
using the model to directly clean its own validation data leads to
underestimates even if total error is low. This indicates a need for
researchers to pursue methods that not only reduce total error but also seek to
de-correlate cleaning error with model scores.
Related papers
- Model Evaluation in the Dark: Robust Classifier Metrics with Missing Labels [2.384873896423002]
We propose a multiple imputation technique for evaluating classifiers using metrics such as precision, recall, and ROC-AUC.
We empirically show that the predictive distribution's location and shape are generally correct, even in the Missing Not At Random regime.
arXiv Detail & Related papers (2025-04-25T14:31:42Z) - Estimating the Conformal Prediction Threshold from Noisy Labels [22.841631892273547]
We show how we can estimate the noise-free conformal threshold based on the noisy labeled data.
We dub our approach Noise-Aware Conformal Prediction (NACP) and show on several natural and medical image classification datasets.
arXiv Detail & Related papers (2025-01-22T09:35:58Z) - Precision and Recall Reject Curves for Classification [1.2507543279181126]
We propose reject curves that evaluate precision and recall, the recall-reject curve and the precision-reject curve.
We show on imbalanced benchmarks and medical, real-world data that for these scenarios, the proposed precision- and recall-curves yield more accurate insights.
arXiv Detail & Related papers (2023-08-16T14:09:48Z) - Label Noise: Correcting the Forward-Correction [0.0]
Training neural network classifiers on datasets with label noise poses a risk of overfitting them to the noisy labels.
We propose an approach to tackling overfitting caused by label noise.
Motivated by this observation, we propose imposing a lower bound on the training loss to mitigate overfitting.
arXiv Detail & Related papers (2023-07-24T19:41:19Z) - Conservative Prediction via Data-Driven Confidence Minimization [70.93946578046003]
In safety-critical applications of machine learning, it is often desirable for a model to be conservative.
We propose the Data-Driven Confidence Minimization framework, which minimizes confidence on an uncertainty dataset.
arXiv Detail & Related papers (2023-06-08T07:05:36Z) - Knockoffs-SPR: Clean Sample Selection in Learning with Noisy Labels [56.81761908354718]
We propose a novel theoretically guaranteed clean sample selection framework for learning with noisy labels.
Knockoffs-SPR can be regarded as a sample selection module for a standard supervised training pipeline.
We further combine it with a semi-supervised algorithm to exploit the support of noisy data as unlabeled data.
arXiv Detail & Related papers (2023-01-02T07:13:28Z) - Is the Performance of My Deep Network Too Good to Be True? A Direct
Approach to Estimating the Bayes Error in Binary Classification [86.32752788233913]
In classification problems, the Bayes error can be used as a criterion to evaluate classifiers with state-of-the-art performance.
We propose a simple and direct Bayes error estimator, where we just take the mean of the labels that show emphuncertainty of the classes.
Our flexible approach enables us to perform Bayes error estimation even for weakly supervised data.
arXiv Detail & Related papers (2022-02-01T13:22:26Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Low-Shot Validation: Active Importance Sampling for Estimating
Classifier Performance on Rare Categories [47.050853657721596]
For machine learning models trained with limited labeled training data, validation stands to become the main bottleneck to reducing overall annotation costs.
We propose a statistical validation algorithm that accurately estimates the F-score of binary classifiers for rare categories.
In particular, we can estimate model F1 scores with a variance of 0.005 using as few as 100 labels.
arXiv Detail & Related papers (2021-09-13T06:01:16Z) - Robustness of Accuracy Metric and its Inspirations in Learning with
Noisy Labels [51.66448070984615]
We show that maximizing training accuracy on sufficiently many noisy samples yields an approximately optimal classifier.
For validation, we prove that a noisy validation set is reliable, addressing the critical demand of model selection.
We show characterizations of models trained with noisy labels, motivated by our theoretical results, and verify the utility of a noisy validation set.
arXiv Detail & Related papers (2020-12-08T03:37:47Z) - Improving Generalization of Deep Fault Detection Models in the Presence
of Mislabeled Data [1.3535770763481902]
We propose a novel two-step framework for robust training with label noise.
In the first step, we identify outliers (including the mislabeled samples) based on the update in the hypothesis space.
In the second step, we propose different approaches to modifying the training data based on the identified outliers and a data augmentation technique.
arXiv Detail & Related papers (2020-09-30T12:33:25Z) - Salvage Reusable Samples from Noisy Data for Robust Learning [70.48919625304]
We propose a reusable sample selection and correction approach, termed as CRSSC, for coping with label noise in training deep FG models with web images.
Our key idea is to additionally identify and correct reusable samples, and then leverage them together with clean examples to update the networks.
arXiv Detail & Related papers (2020-08-06T02:07:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.