Decoupled Contrastive Learning for Long-Tailed Recognition
- URL: http://arxiv.org/abs/2403.06151v1
- Date: Sun, 10 Mar 2024 09:46:28 GMT
- Title: Decoupled Contrastive Learning for Long-Tailed Recognition
- Authors: Shiyu Xuan, Shiliang Zhang
- Abstract summary: Supervised Contrastive Loss (SCL) is popular in visual representation learning.
In the scenario of long-tailed recognition, where the number of samples in each class is imbalanced, treating two types of positive samples equally leads to the biased optimization for intra-category distance.
We propose a patch-based self distillation to transfer knowledge from head to tail classes to relieve the under-representation of tail classes.
- Score: 58.255966442426484
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Supervised Contrastive Loss (SCL) is popular in visual representation
learning. Given an anchor image, SCL pulls two types of positive samples, i.e.,
its augmentation and other images from the same class together, while pushes
negative images apart to optimize the learned embedding. In the scenario of
long-tailed recognition, where the number of samples in each class is
imbalanced, treating two types of positive samples equally leads to the biased
optimization for intra-category distance. In addition, similarity relationship
among negative samples, that are ignored by SCL, also presents meaningful
semantic cues. To improve the performance on long-tailed recognition, this
paper addresses those two issues of SCL by decoupling the training objective.
Specifically, it decouples two types of positives in SCL and optimizes their
relations toward different objectives to alleviate the influence of the
imbalanced dataset. We further propose a patch-based self distillation to
transfer knowledge from head to tail classes to relieve the
under-representation of tail classes. It uses patch-based features to mine
shared visual patterns among different instances and leverages a self
distillation procedure to transfer such knowledge. Experiments on different
long-tailed classification benchmarks demonstrate the superiority of our
method. For instance, it achieves the 57.7% top-1 accuracy on the ImageNet-LT
dataset. Combined with the ensemble-based method, the performance can be
further boosted to 59.7%, which substantially outperforms many recent works.
The code is available at https://github.com/SY-Xuan/DSCL.
Related papers
- Rank Supervised Contrastive Learning for Time Series Classification [17.302643963704643]
We present Rank Supervised Contrastive Learning (RankSCL) to perform time series classification.
RankSCL augments raw data in a targeted way in the embedding space.
A novel rank loss is developed to assign different weights for different levels of positive samples.
arXiv Detail & Related papers (2024-01-31T18:29:10Z) - Balanced Contrastive Learning for Long-Tailed Visual Recognition [32.789465918318925]
Real-world data typically follow a long-tailed distribution, where a few majority categories occupy most of the data.
In this paper, we focus on representation learning for imbalanced data.
We propose a novel loss for balanced contrastive learning (BCL)
arXiv Detail & Related papers (2022-07-19T03:48:59Z) - Decoupled Contrastive Learning [23.25775900388382]
We identify a noticeable negative-positive-coupling (NPC) effect in the widely used cross-entropy (InfoNCE) loss.
By properly addressing the NPC effect, we reach a decoupled contrastive learning (DCL) objective function.
Our approach achieves $66.9%$ ImageNet top-1 accuracy using batch size 256 within 200 epochs pre-training, outperforming its baseline SimCLR by $5.1%$.
arXiv Detail & Related papers (2021-10-13T16:38:43Z) - Semi-supervised Contrastive Learning with Similarity Co-calibration [72.38187308270135]
We propose a novel training strategy, termed as Semi-supervised Contrastive Learning (SsCL)
SsCL combines the well-known contrastive loss in self-supervised learning with the cross entropy loss in semi-supervised learning.
We show that SsCL produces more discriminative representation and is beneficial to few shot learning.
arXiv Detail & Related papers (2021-05-16T09:13:56Z) - Contrastive Attraction and Contrastive Repulsion for Representation
Learning [131.72147978462348]
Contrastive learning (CL) methods learn data representations in a self-supervision manner, where the encoder contrasts each positive sample over multiple negative samples.
Recent CL methods have achieved promising results when pretrained on large-scale datasets, such as ImageNet.
We propose a doubly CL strategy that separately compares positive and negative samples within their own groups, and then proceeds with a contrast between positive and negative groups.
arXiv Detail & Related papers (2021-05-08T17:25:08Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - Contrastive Learning with Adversarial Examples [79.39156814887133]
Contrastive learning (CL) is a popular technique for self-supervised learning (SSL) of visual representations.
This paper introduces a new family of adversarial examples for constrastive learning and using these examples to define a new adversarial training algorithm for SSL, denoted as CLAE.
arXiv Detail & Related papers (2020-10-22T20:45:10Z) - Unsupervised Feature Learning by Cross-Level Instance-Group
Discrimination [68.83098015578874]
We integrate between-instance similarity into contrastive learning, not directly by instance grouping, but by cross-level discrimination.
CLD effectively brings unsupervised learning closer to natural data and real-world applications.
New state-of-the-art on self-supervision, semi-supervision, and transfer learning benchmarks, and beats MoCo v2 and SimCLR on every reported performance.
arXiv Detail & Related papers (2020-08-09T21:13:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.