Class Interference Regularization
- URL: http://arxiv.org/abs/2009.02396v1
- Date: Fri, 4 Sep 2020 21:03:32 GMT
- Title: Class Interference Regularization
- Authors: Bharti Munjal, Sikandar Amin, Fabio Galasso
- Abstract summary: Contrastive losses yield state-of-the-art performance for person re-identification, face verification and few shot learning.
We propose a novel, simple and effective regularization technique, the Class Interference Regularization (CIR)
CIR perturbs the output features by randomly moving them towards the average embeddings of the negative classes.
- Score: 7.248447600071719
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive losses yield state-of-the-art performance for person
re-identification, face verification and few shot learning. They have recently
outperformed the cross-entropy loss on classification at the ImageNet scale and
outperformed all self-supervision prior results by a large margin (SimCLR).
Simple and effective regularization techniques such as label smoothing and
self-distillation do not apply anymore, because they act on multinomial label
distributions, adopted in cross-entropy losses, and not on tuple comparative
terms, which characterize the contrastive losses.
Here we propose a novel, simple and effective regularization technique, the
Class Interference Regularization (CIR), which applies to cross-entropy losses
but is especially effective on contrastive losses. CIR perturbs the output
features by randomly moving them towards the average embeddings of the negative
classes. To the best of our knowledge, CIR is the first regularization
technique to act on the output features.
In experimental evaluation, the combination of CIR and a plain Siamese-net
with triplet loss yields best few-shot learning performance on the challenging
tieredImageNet. CIR also improves the state-of-the-art technique in person
re-identification on the Market-1501 dataset, based on triplet loss, and the
state-of-the-art technique in person search on the CUHK-SYSU dataset, based on
a cross-entropy loss. Finally, on the task of classification CIR performs on
par with the popular label smoothing, as demonstrated for CIFAR-10 and -100.
Related papers
- A Unified Contrastive Loss for Self-Training [3.3454373538792552]
Self-training methods have proven to be effective in exploiting abundant unlabeled data in semi-supervised learning.
We propose a general framework to enhance self-training methods, which replaces all instances of CE losses with a unique contrastive loss.
Our framework results in significant performance improvements across three different datasets with a limited number of labeled data.
arXiv Detail & Related papers (2024-09-11T14:22:41Z) - EnsLoss: Stochastic Calibrated Loss Ensembles for Preventing Overfitting in Classification [1.3778851745408134]
We propose a novel ensemble method, namely EnsLoss, to combine loss functions within the Empirical risk minimization framework.
We first transform the CC conditions of losses into loss-derivatives, thereby bypassing the need for explicit loss functions.
We theoretically establish the statistical consistency of our approach and provide insights into its benefits.
arXiv Detail & Related papers (2024-09-02T02:40:42Z) - Bayesian Learning-driven Prototypical Contrastive Loss for Class-Incremental Learning [42.14439854721613]
We propose a prototypical network with a Bayesian learning-driven contrastive loss (BLCL) tailored specifically for class-incremental learning scenarios.
Our approach dynamically adapts the balance between the cross-entropy and contrastive loss functions with a Bayesian learning technique.
arXiv Detail & Related papers (2024-05-17T19:49:02Z) - CR-SAM: Curvature Regularized Sharpness-Aware Minimization [8.248964912483912]
Sharpness-Aware Minimization (SAM) aims to enhance the generalizability by minimizing worst-case loss using one-step gradient ascent as an approximation.
In this paper, we introduce a normalized Hessian trace to accurately measure the curvature of loss landscape on em both training and test sets.
In particular, to counter excessive non-linearity of loss landscape, we propose Curvature Regularized SAM (CR-SAM)
arXiv Detail & Related papers (2023-12-21T03:46:29Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - Tuned Contrastive Learning [77.67209954169593]
We propose a novel contrastive loss function -- Tuned Contrastive Learning (TCL) loss.
TCL generalizes to multiple positives and negatives in a batch and offers parameters to tune and improve the gradient responses from hard positives and hard negatives.
We show how to extend TCL to self-supervised setting and empirically compare it with various SOTA self-supervised learning methods.
arXiv Detail & Related papers (2023-05-18T03:26:37Z) - Do Lessons from Metric Learning Generalize to Image-Caption Retrieval? [67.45267657995748]
The triplet loss with semi-hard negatives has become the de facto choice for image-caption retrieval (ICR) methods that are optimized from scratch.
Recent progress in metric learning has given rise to new loss functions that outperform the triplet loss on tasks such as image retrieval and representation learning.
We ask whether these findings generalize to the setting of ICR by comparing three loss functions on two ICR methods.
arXiv Detail & Related papers (2022-02-14T15:18:00Z) - Label Distributionally Robust Losses for Multi-class Classification:
Consistency, Robustness and Adaptivity [55.29408396918968]
We study a family of loss functions named label-distributionally robust (LDR) losses for multi-class classification.
Our contributions include both consistency and robustness by establishing top-$k$ consistency of LDR losses for multi-class classification.
We propose a new adaptive LDR loss that automatically adapts the individualized temperature parameter to the noise degree of class label of each instance.
arXiv Detail & Related papers (2021-12-30T00:27:30Z) - Orthogonal Projection Loss [59.61277381836491]
We develop a novel loss function termed Orthogonal Projection Loss' (OPL)
OPL directly enforces inter-class separation alongside intra-class clustering in the feature space.
OPL offers unique advantages as it does not require careful negative mining and is not sensitive to the batch size.
arXiv Detail & Related papers (2021-03-25T17:58:00Z) - A Flatter Loss for Bias Mitigation in Cross-dataset Facial Age
Estimation [37.107335288543624]
We advocate a cross-dataset protocol for age estimation benchmarking.
We propose a novel loss function that is more effective for neural network training.
arXiv Detail & Related papers (2020-10-20T15:22:29Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.