Rediscovering BCE Loss for Uniform Classification
- URL: http://arxiv.org/abs/2403.07289v1
- Date: Tue, 12 Mar 2024 03:44:40 GMT
- Title: Rediscovering BCE Loss for Uniform Classification
- Authors: Qiufu Li, Xi Jia, Jiancan Zhou, Linlin Shen, Jinming Duan
- Abstract summary: This paper introduces the concept of uniform classification, which employs a unified threshold to classify all samples.
We propose the uniform classification accuracy as a metric to measure the model's performance in uniform classification.
- Score: 35.66000285310775
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces the concept of uniform classification, which employs a
unified threshold to classify all samples rather than adaptive threshold
classifying each individual sample. We also propose the uniform classification
accuracy as a metric to measure the model's performance in uniform
classification. Furthermore, begin with a naive loss, we mathematically derive
a loss function suitable for the uniform classification, which is the BCE
function integrated with a unified bias. We demonstrate the unified threshold
could be learned via the bias. The extensive experiments on six classification
datasets and three feature extraction models show that, compared to the SoftMax
loss, the models trained with the BCE loss not only exhibit higher uniform
classification accuracy but also higher sample-wise classification accuracy. In
addition, the learned bias from BCE loss is very close to the unified threshold
used in the uniform classification. The features extracted by the models
trained with BCE loss not only possess uniformity but also demonstrate better
intra-class compactness and inter-class distinctiveness, yielding superior
performance on open-set tasks such as face recognition.
Related papers
- BCE vs. CE in Deep Feature Learning [33.24161955363104]
We compare binary CE (BCE) and cross-entropy (CE) in deep feature learning.<n>BCE can also maximize the intra-class compactness and inter-class distinctiveness when reaching its minimum.<n>BCE measures the absolute values of decision scores and adjust the positive/negative decision scores across all samples to uniformly high/low levels.
arXiv Detail & Related papers (2025-05-09T06:18:31Z) - Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - Latent Enhancing AutoEncoder for Occluded Image Classification [2.6217304977339473]
We introduce LEARN: Latent Enhancing feAture Reconstruction Network.
An auto-encoder based network that can be incorporated into the classification model before its head.
On the OccludedPASCAL3D+ dataset, the proposed LEARN outperforms standard classification models.
arXiv Detail & Related papers (2024-02-10T12:22:31Z) - Uncertainty-guided Boundary Learning for Imbalanced Social Event
Detection [64.4350027428928]
We propose a novel uncertainty-guided class imbalance learning framework for imbalanced social event detection tasks.
Our model significantly improves social event representation and classification tasks in almost all classes, especially those uncertain ones.
arXiv Detail & Related papers (2023-10-30T03:32:04Z) - Bias Amplification Enhances Minority Group Performance [10.380812738348899]
We propose BAM, a novel two-stage training algorithm.
In the first stage, the model is trained using a bias amplification scheme via introducing a learnable auxiliary variable for each training sample.
In the second stage, we upweight the samples that the bias-amplified model misclassifies, and then continue training the same model on the reweighted dataset.
arXiv Detail & Related papers (2023-09-13T04:40:08Z) - A Visual Interpretation-Based Self-Improved Classification System Using
Virtual Adversarial Training [4.722922834127293]
This paper proposes a visual interpretation-based self-improving classification model with a combination of virtual adversarial training (VAT) and BERT models to address the problems.
Specifically, a fine-tuned BERT model is used as a classifier to classify the sentiment of the text.
The predicted sentiment classification labels are used as part of the input of another BERT for spam classification via a semi-supervised training manner.
arXiv Detail & Related papers (2023-09-03T15:07:24Z) - Generative Robust Classification [3.4773470589069477]
Training adversarially robust discriminative (i.e., softmax) classification has been the dominant approach to robust classification.
We investigate using adversarial training (AT)-based generative models.
We find it straightforward to apply advanced data augmentation to achieve better robustness in our approach.
arXiv Detail & Related papers (2022-12-14T15:33:11Z) - Score-Based Generative Classifiers [9.063815952852783]
Generative models have been used as adversarially robust classifiers on simple datasets such as MNIST.
Previous results have suggested a trade-off between the likelihood of the data and classification accuracy.
We show that score-based generative models are closing the gap in classification accuracy compared to standard discriminative models.
arXiv Detail & Related papers (2021-10-01T15:05:33Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.