Improving Screening Processes via Calibrated Subset Selection
- URL: http://arxiv.org/abs/2202.01147v3
- Date: Mon, 13 Jun 2022 01:02:35 GMT
- Title: Improving Screening Processes via Calibrated Subset Selection
- Authors: Lequn Wang, Thorsten Joachims, Manuel Gomez Rodriguez
- Abstract summary: We develop a distribution-free screening algorithm called Calibrated Subset Selection (CSS)
CSS finds near-optimal shortlists of candidates that contain a desired number of qualified candidates in expectation.
Experiments on US Census survey data validate our theoretical results and show that the shortlists provided by our algorithm are superior to those provided by several competitive baselines.
- Score: 35.952153033163576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many selection processes such as finding patients qualifying for a medical
trial or retrieval pipelines in search engines consist of multiple stages,
where an initial screening stage focuses the resources on shortlisting the most
promising candidates. In this paper, we investigate what guarantees a screening
classifier can provide, independently of whether it is constructed manually or
trained. We find that current solutions do not enjoy distribution-free
theoretical guarantees -- we show that, in general, even for a perfectly
calibrated classifier, there always exist specific pools of candidates for
which its shortlist is suboptimal. Then, we develop a distribution-free
screening algorithm -- called Calibrated Subset Selection (CSS) -- that, given
any classifier and some amount of calibration data, finds near-optimal
shortlists of candidates that contain a desired number of qualified candidates
in expectation. Moreover, we show that a variant of CSS that calibrates a given
classifier multiple times across specific groups can create shortlists with
provable diversity guarantees. Experiments on US Census survey data validate
our theoretical results and show that the shortlists provided by our algorithm
are superior to those provided by several competitive baselines.
Related papers
- Centralized Selection with Preferences in the Presence of Biases [25.725937202777267]
The paper focuses on the setting in which candidates are divided into multiple groups and the observed utilities of candidates in some groups are biased--systematically lower than their true utilities.
An algorithm is presented along with proof that it produces selections that achieve near-optimal group fairness with respect to preferences while also nearly maximizing the true utility under distributional assumptions.
arXiv Detail & Related papers (2024-09-07T19:47:13Z) - Multi-Candidate Speculative Decoding [82.05519287513444]
Large language models have shown impressive capabilities across a variety of NLP tasks, yet their generating text autoregressively is time-consuming.
One way to speed them up is speculative decoding, which generates candidate segments from a fast draft model that is then verified in parallel by the target model.
This paper proposes sampling multiple candidates from a draft model and then organising them in batches for verification.
We design algorithms for efficient multi-candidate verification while maintaining the distribution of the target model.
arXiv Detail & Related papers (2024-01-12T17:15:23Z) - Large Language Models Are Not Robust Multiple Choice Selectors [117.72712117510953]
Multiple choice questions (MCQs) serve as a common yet important task format in the evaluation of large language models (LLMs)
This work shows that modern LLMs are vulnerable to option position changes due to their inherent "selection bias"
We propose a label-free, inference-time debiasing method, called PriDe, which separates the model's prior bias for option IDs from the overall prediction distribution.
arXiv Detail & Related papers (2023-09-07T17:44:56Z) - Analysis of the Cambridge Multiple-Choice Questions Reading Dataset with
a Focus on Candidate Response Distribution [38.58190457533888]
We introduce the task of candidate distribution matching, propose several evaluation metrics for the task, and demonstrate that automatic systems trained on RACE++ can be leveraged as baselines for our task.
We further demonstrate that these automatic systems can be used for practical pre-test evaluation tasks such as detecting underperforming distractors.
arXiv Detail & Related papers (2023-06-22T17:13:08Z) - Revisiting Long-tailed Image Classification: Survey and Benchmarks with
New Evaluation Metrics [88.39382177059747]
A corpus of metrics is designed for measuring the accuracy, robustness, and bounds of algorithms for learning with long-tailed distribution.
Based on our benchmarks, we re-evaluate the performance of existing methods on CIFAR10 and CIFAR100 datasets.
arXiv Detail & Related papers (2023-02-03T02:40:54Z) - On the Within-Group Fairness of Screening Classifiers [16.404065044314976]
We argue that screening policies that use calibrated classifiers may suffer from an understudied type of within-group unfairness.
We show that within-group monotonicity can be achieved at a small cost in terms of prediction granularity and shortlist size.
arXiv Detail & Related papers (2023-01-31T19:00:28Z) - Selection by Prediction with Conformal p-values [7.917044695538599]
We study screening procedures that aim to select candidates whose unobserved outcomes exceed user-specified values.
We develop a method that wraps around any prediction model to produce a subset of candidates while controlling the proportion of falsely selected units.
arXiv Detail & Related papers (2022-10-04T06:34:49Z) - Ensemble pruning via an integer programming approach with diversity
constraints [0.0]
In this paper, we consider a binary classification problem and propose an integer programming (IP) approach for selecting optimal subsets.
We also propose constraints to ensure minimum diversity levels in the ensemble.
Our approach yields competitive results when compared to some of the best and most used pruning methods in literature.
arXiv Detail & Related papers (2022-05-02T17:59:11Z) - Online Active Model Selection for Pre-trained Classifiers [72.84853880948894]
We design an online selective sampling approach that actively selects informative examples to label and outputs the best model with high probability at any round.
Our algorithm can be used for online prediction tasks for both adversarial and streams.
arXiv Detail & Related papers (2020-10-19T19:53:15Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.