Classification with Strategically Withheld Data
- URL: http://arxiv.org/abs/2012.10203v2
- Date: Thu, 14 Jan 2021 12:13:21 GMT
- Title: Classification with Strategically Withheld Data
- Authors: Anilesh K. Krishnaswamy, Haoming Li, David Rein, Hanrui Zhang, and
Vincent Conitzer
- Abstract summary: Machine learning techniques can be useful in applications such as credit approval and college admission.
To be classified more favorably in such contexts, an agent may decide to strategically withhold some of her features, such as bad test scores.
We design three classification methods: sc Mincut, sc Hill-Climbing (sc HC) and Incentive- Logistic Regression (sc-LR)
- Score: 41.78264347024645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning techniques can be useful in applications such as credit
approval and college admission. However, to be classified more favorably in
such contexts, an agent may decide to strategically withhold some of her
features, such as bad test scores. This is a missing data problem with a twist:
which data is missing {\em depends on the chosen classifier}, because the
specific classifier is what may create the incentive to withhold certain
feature values. We address the problem of training classifiers that are robust
to this behavior.
We design three classification methods: {\sc Mincut}, {\sc Hill-Climbing}
({\sc HC}) and Incentive-Compatible Logistic Regression ({\sc IC-LR}). We show
that {\sc Mincut} is optimal when the true distribution of data is fully known.
However, it can produce complex decision boundaries, and hence be prone to
overfitting in some cases. Based on a characterization of truthful classifiers
(i.e., those that give no incentive to strategically hide features), we devise
a simpler alternative called {\sc HC} which consists of a hierarchical ensemble
of out-of-the-box classifiers, trained using a specialized hill-climbing
procedure which we show to be convergent. For several reasons, {\sc Mincut} and
{\sc HC} are not effective in utilizing a large number of complementarily
informative features. To this end, we present {\sc IC-LR}, a modification of
Logistic Regression that removes the incentive to strategically drop features.
We also show that our algorithms perform well in experiments on real-world data
sets, and present insights into their relative performance in different
settings.
Related papers
- A Hard-to-Beat Baseline for Training-free CLIP-based Adaptation [121.0693322732454]
Contrastive Language-Image Pretraining (CLIP) has gained popularity for its remarkable zero-shot capacity.
Recent research has focused on developing efficient fine-tuning methods to enhance CLIP's performance in downstream tasks.
We revisit a classical algorithm, Gaussian Discriminant Analysis (GDA), and apply it to the downstream classification of CLIP.
arXiv Detail & Related papers (2024-02-06T15:45:27Z) - SSB: Simple but Strong Baseline for Boosting Performance of Open-Set
Semi-Supervised Learning [106.46648817126984]
In this paper, we study the challenging and realistic open-set SSL setting.
The goal is to both correctly classify inliers and to detect outliers.
We find that inlier classification performance can be largely improved by incorporating high-confidence pseudo-labeled data.
arXiv Detail & Related papers (2023-11-17T15:14:40Z) - An Upper Bound for the Distribution Overlap Index and Its Applications [18.481370450591317]
This paper proposes an easy-to-compute upper bound for the overlap index between two probability distributions.
The proposed bound shows its value in one-class classification and domain shift analysis.
Our work shows significant promise toward broadening the applications of overlap-based metrics.
arXiv Detail & Related papers (2022-12-16T20:02:03Z) - Class-Level Logit Perturbation [0.0]
Feature perturbation and label perturbation have been proven to be useful in various deep learning approaches.
New methodologies are proposed to explicitly learn to perturb logits for both single-label and multi-label classification tasks.
As it only perturbs on logit, it can be used as a plug-in to fuse with any existing classification algorithms.
arXiv Detail & Related papers (2022-09-13T00:49:32Z) - A Study on the Predictability of Sample Learning Consistency [4.022364531869169]
We train models to predict C-Score for CIFAR-100 and CIFAR-10.
We find, however, that these models generalize poorly both within the same distribution as well as out of distribution.
We hypothesize that a sample's relation to its neighbours, in particular, how many of them share the same labels, can help in explaining C-Scores.
arXiv Detail & Related papers (2022-07-07T21:05:53Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Exploring Category-correlated Feature for Few-shot Image Classification [27.13708881431794]
We present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
The proposed approach consistently obtains considerable performance gains on three widely used benchmarks.
arXiv Detail & Related papers (2021-12-14T08:25:24Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - For self-supervised learning, Rationality implies generalization,
provably [13.526562756159809]
We prove a new upper bound on the generalization gap of classifiers obtained by first using self-supervision.
We show that our bound is non-vacuous for many popular representation-learning based classifiers on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2020-10-16T17:07:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.