Improving Dense Contrastive Learning with Dense Negative Pairs
- URL: http://arxiv.org/abs/2210.05063v1
- Date: Tue, 11 Oct 2022 00:26:59 GMT
- Title: Improving Dense Contrastive Learning with Dense Negative Pairs
- Authors: Berk Iskender, Zhenlin Xu, Simon Kornblith, Enhung Chu, Maryam Khademi
- Abstract summary: We study how to improve the quality of the representations learned by DenseCL by modifying the training scheme and objective function.
Our results show 3.5% and 4% mAP improvement over SimCLR and DenseCL in COCO multi-label classification.
- Score: 15.728417548134047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many contrastive representation learning methods learn a single global
representation of an entire image. However, dense contrastive representation
learning methods such as DenseCL [19] can learn better representations for
tasks requiring stronger spatial localization of features, such as multi-label
classification, detection, and segmentation. In this work, we study how to
improve the quality of the representations learned by DenseCL by modifying the
training scheme and objective function, and propose DenseCL++. We also conduct
several ablation studies to better understand the effects of: (i) various
techniques to form dense negative pairs among augmentations of different
images, (ii) cross-view dense negative and positive pairs, and (iii) an
auxiliary reconstruction task. Our results show 3.5% and 4% mAP improvement
over SimCLR [3] and DenseCL in COCO multi-label classification. In COCO and VOC
segmentation tasks, we achieve 1.8% and 0.7% mIoU improvements over SimCLR,
respectively.
Related papers
- Synergy and Diversity in CLIP: Enhancing Performance Through Adaptive Backbone Ensembling [58.50618448027103]
Contrastive Language-Image Pretraining (CLIP) stands out as a prominent method for image representation learning.
This paper explores the differences across various CLIP-trained vision backbones.
Method achieves a remarkable increase in accuracy of up to 39.1% over the best single backbone.
arXiv Detail & Related papers (2024-05-27T12:59:35Z) - LeOCLR: Leveraging Original Images for Contrastive Learning of Visual Representations [4.680881326162484]
Contrastive instance discrimination methods outperform supervised learning in downstream tasks such as image classification and object detection.
A common augmentation technique in contrastive learning is random cropping followed by resizing.
We introduce LeOCLR, a framework that employs a novel instance discrimination approach and an adapted loss function.
arXiv Detail & Related papers (2024-03-11T15:33:32Z) - Decoupled Contrastive Learning for Long-Tailed Recognition [58.255966442426484]
Supervised Contrastive Loss (SCL) is popular in visual representation learning.
In the scenario of long-tailed recognition, where the number of samples in each class is imbalanced, treating two types of positive samples equally leads to the biased optimization for intra-category distance.
We propose a patch-based self distillation to transfer knowledge from head to tail classes to relieve the under-representation of tail classes.
arXiv Detail & Related papers (2024-03-10T09:46:28Z) - PARTICLE: Part Discovery and Contrastive Learning for Fine-grained
Recognition [29.05343410791649]
We develop techniques for refining representations for fine-grained classification and segmentation tasks in a self-supervised manner.
We show that a part-centric learning step aggregates and contrasts representations of parts within an image.
This improves the performance on image classification and part segmentation tasks across datasets.
arXiv Detail & Related papers (2023-09-25T02:08:48Z) - Hallucination Improves the Performance of Unsupervised Visual
Representation Learning [9.504503675097137]
We propose Hallucinator that could efficiently generate additional positive samples for further contrast.
The Hallucinator is differentiable and creates new data in the feature space.
Remarkably, we empirically prove that the proposed Hallucinator generalizes well to various contrastive learning models.
arXiv Detail & Related papers (2023-07-22T21:15:56Z) - Semantic Positive Pairs for Enhancing Visual Representation Learning of Instance Discrimination methods [4.680881326162484]
Self-supervised learning algorithms (SSL) based on instance discrimination have shown promising results.
We propose an approach to identify those images with similar semantic content and treat them as positive instances.
We run experiments on three benchmark datasets: ImageNet, STL-10 and CIFAR-10 with different instance discrimination SSL approaches.
arXiv Detail & Related papers (2023-06-28T11:47:08Z) - Rethinking Semi-Supervised Medical Image Segmentation: A
Variance-Reduction Perspective [51.70661197256033]
We propose ARCO, a semi-supervised contrastive learning framework with stratified group theory for medical image segmentation.
We first propose building ARCO through the concept of variance-reduced estimation and show that certain variance-reduction techniques are particularly beneficial in pixel/voxel-level segmentation tasks.
We experimentally validate our approaches on eight benchmarks, i.e., five 2D/3D medical and three semantic segmentation datasets, with different label settings.
arXiv Detail & Related papers (2023-02-03T13:50:25Z) - Provable Stochastic Optimization for Global Contrastive Learning: Small
Batch Does Not Harm Performance [53.49803579981569]
We consider a global objective for contrastive learning, which contrasts each positive pair with all negative pairs for an anchor point.
Existing methods such as SimCLR requires a large batch size in order to achieve a satisfactory result.
We propose a memory-efficient optimization algorithm for solving the Global Contrastive Learning of Representations, named SogCLR.
arXiv Detail & Related papers (2022-02-24T22:16:53Z) - Dense Contrastive Learning for Self-Supervised Visual Pre-Training [102.15325936477362]
We present dense contrastive learning, which implements self-supervised learning by optimizing a pairwise contrastive (dis)similarity loss at the pixel level between two views of input images.
Compared to the baseline method MoCo-v2, our method introduces negligible computation overhead (only 1% slower)
arXiv Detail & Related papers (2020-11-18T08:42:32Z) - Contrastive Clustering [57.71729650297379]
We propose Contrastive Clustering (CC) which explicitly performs the instance- and cluster-level contrastive learning.
In particular, CC achieves an NMI of 0.705 (0.431) on the CIFAR-10 (CIFAR-100) dataset, which is an up to 19% (39%) performance improvement compared with the best baseline.
arXiv Detail & Related papers (2020-09-21T08:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.