Semi-supervised Contrastive Learning with Similarity Co-calibration
- URL: http://arxiv.org/abs/2105.07387v1
- Date: Sun, 16 May 2021 09:13:56 GMT
- Title: Semi-supervised Contrastive Learning with Similarity Co-calibration
- Authors: Yuhang Zhang and Xiaopeng Zhang and Robert.C.Qiu and Jie Li and
Haohang Xu and Qi Tian
- Abstract summary: We propose a novel training strategy, termed as Semi-supervised Contrastive Learning (SsCL)
SsCL combines the well-known contrastive loss in self-supervised learning with the cross entropy loss in semi-supervised learning.
We show that SsCL produces more discriminative representation and is beneficial to few shot learning.
- Score: 72.38187308270135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semi-supervised learning acts as an effective way to leverage massive
unlabeled data. In this paper, we propose a novel training strategy, termed as
Semi-supervised Contrastive Learning (SsCL), which combines the well-known
contrastive loss in self-supervised learning with the cross entropy loss in
semi-supervised learning, and jointly optimizes the two objectives in an
end-to-end way. The highlight is that different from self-training based
semi-supervised learning that conducts prediction and retraining over the same
model weights, SsCL interchanges the predictions over the unlabeled data
between the two branches, and thus formulates a co-calibration procedure, which
we find is beneficial for better prediction and avoid being trapped in local
minimum. Towards this goal, the contrastive loss branch models pairwise
similarities among samples, using the nearest neighborhood generated from the
cross entropy branch, and in turn calibrates the prediction distribution of the
cross entropy branch with the contrastive similarity. We show that SsCL
produces more discriminative representation and is beneficial to few shot
learning. Notably, on ImageNet with ResNet50 as the backbone, SsCL achieves
60.2% and 72.1% top-1 accuracy with 1% and 10% labeled samples, respectively,
which significantly outperforms the baseline, and is better than previous
semi-supervised and self-supervised methods.
Related papers
- Bayesian Learning-driven Prototypical Contrastive Loss for Class-Incremental Learning [42.14439854721613]
We propose a prototypical network with a Bayesian learning-driven contrastive loss (BLCL) tailored specifically for class-incremental learning scenarios.
Our approach dynamically adapts the balance between the cross-entropy and contrastive loss functions with a Bayesian learning technique.
arXiv Detail & Related papers (2024-05-17T19:49:02Z) - Decoupled Contrastive Learning for Long-Tailed Recognition [58.255966442426484]
Supervised Contrastive Loss (SCL) is popular in visual representation learning.
In the scenario of long-tailed recognition, where the number of samples in each class is imbalanced, treating two types of positive samples equally leads to the biased optimization for intra-category distance.
We propose a patch-based self distillation to transfer knowledge from head to tail classes to relieve the under-representation of tail classes.
arXiv Detail & Related papers (2024-03-10T09:46:28Z) - Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation [63.180725016463974]
Cross-modal retrieval relies on well-matched large-scale datasets that are laborious in practice.
We introduce a novel noisy correspondence learning framework, namely textbfSelf-textbfReinforcing textbfErrors textbfMitigation (SREM)
arXiv Detail & Related papers (2023-12-27T09:03:43Z) - Rethinking Prototypical Contrastive Learning through Alignment,
Uniformity and Correlation [24.794022951873156]
We propose to learn Prototypical representation through Alignment, Uniformity and Correlation (PAUC)
Specifically, the ordinary ProtoNCE loss is revised with: (1) an alignment loss that pulls embeddings from positive prototypes together; (2) a loss that distributes the prototypical level features uniformly; (3) a correlation loss that increases the diversity and discriminability between prototypical level features.
arXiv Detail & Related papers (2022-10-18T22:33:12Z) - Hierarchical Semi-Supervised Contrastive Learning for
Contamination-Resistant Anomaly Detection [81.07346419422605]
Anomaly detection aims at identifying deviant samples from the normal data distribution.
Contrastive learning has provided a successful way to sample representation that enables effective discrimination on anomalies.
We propose a novel hierarchical semi-supervised contrastive learning framework, for contamination-resistant anomaly detection.
arXiv Detail & Related papers (2022-07-24T18:49:26Z) - Is it all a cluster game? -- Exploring Out-of-Distribution Detection
based on Clustering in the Embedding Space [7.856998585396422]
It is essential for safety-critical applications of deep neural networks to determine when new inputs are significantly different from the training distribution.
We study the structure and separation of clusters in the embedding space and find that supervised contrastive learning leads to well-separated clusters.
In our analysis of different training methods, clustering strategies, distance metrics, and thresholding approaches, we observe that there is no clear winner.
arXiv Detail & Related papers (2022-03-16T11:22:23Z) - S5CL: Unifying Fully-Supervised, Self-Supervised, and Semi-Supervised
Learning Through Hierarchical Contrastive Learning [0.22940141855172028]
We introduce S5CL, a unified framework for fully-supervised, self-supervised, and semi-supervised learning.
With three contrastive losses defined for labeled, unlabeled, and pseudo-labeled images, S5CL can learn feature representations that reflect the hierarchy of distance relationships.
arXiv Detail & Related papers (2022-03-14T17:10:01Z) - Robust Pre-Training by Adversarial Contrastive Learning [120.33706897927391]
Recent work has shown that, when integrated with adversarial training, self-supervised pre-training can lead to state-of-the-art robustness.
We improve robustness-aware self-supervised pre-training by learning representations consistent under both data augmentations and adversarial perturbations.
arXiv Detail & Related papers (2020-10-26T04:44:43Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.