Contrastive Principal Component Learning: Modeling Similarity by
Augmentation Overlap
- URL: http://arxiv.org/abs/2206.00471v1
- Date: Wed, 1 Jun 2022 13:03:58 GMT
- Title: Contrastive Principal Component Learning: Modeling Similarity by
Augmentation Overlap
- Authors: Lu Han, Han-Jia Ye, De-Chuan Zhan
- Abstract summary: We propose a novel Contrastive Principal Component Learning (CPCL) method composed of a contrastive-like loss and an on-the-fly projection loss.
By CPCL, the learned low-dimensional embeddings theoretically preserve the similarity of augmentation distribution between samples.
- Score: 50.48888534815361
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional self-supervised contrastive learning methods learn embeddings by
pulling views of the same sample together and pushing views of different
samples away. Since views of a sample are usually generated via data
augmentations, the semantic relationship between samples is ignored. Based on
the observation that semantically similar samples are more likely to have
similar augmentations, we propose to measure similarity via the distribution of
augmentations, i.e., how much the augmentations of two samples overlap. To
handle the dimensional and computational complexity, we propose a novel
Contrastive Principal Component Learning (CPCL) method composed of a
contrastive-like loss and an on-the-fly projection loss to efficiently perform
PCA on the augmentation feature, which encodes the augmentation distribution.
By CPCL, the learned low-dimensional embeddings theoretically preserve the
similarity of augmentation distribution between samples. Empirical results show
our method can achieve competitive results against various traditional
contrastive learning methods on different benchmarks.
Related papers
- CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective [48.99488315273868]
We present a contrastive knowledge distillation approach, which can be formulated as a sample-wise alignment problem with intra- and inter-sample constraints.
Our method minimizes logit differences within the same sample by considering their numerical values.
We conduct comprehensive experiments on three datasets including CIFAR-100, ImageNet-1K, and MS COCO.
arXiv Detail & Related papers (2024-04-22T11:52:40Z) - Synthetic Hard Negative Samples for Contrastive Learning [8.776888865665024]
This paper proposes a novel feature-level method, namely sampling synthetic hard negative samples for contrastive learning (SSCL)
We generate more and harder negative samples by mixing negative samples, and then sample them by controlling the contrast of anchor sample with the other negative samples.
Our proposed method improves the classification performance on different image datasets and can be readily integrated into existing methods.
arXiv Detail & Related papers (2023-04-06T09:54:35Z) - Hybrid Contrastive Constraints for Multi-Scenario Ad Ranking [38.666592866591344]
Multi-scenario ad ranking aims at leveraging the data from multiple domains or channels for training a unified ranking model.
We propose a Hybrid Contrastive Constrained approach (HC2) for multi-scenario ad ranking.
arXiv Detail & Related papers (2023-02-06T09:15:39Z) - Beyond Instance Discrimination: Relation-aware Contrastive
Self-supervised Learning [75.46664770669949]
We present relation-aware contrastive self-supervised learning (ReCo) to integrate instance relations.
Our ReCo consistently gains remarkable performance improvements.
arXiv Detail & Related papers (2022-11-02T03:25:28Z) - R\'enyiCL: Contrastive Representation Learning with Skew R\'enyi
Divergence [78.15455360335925]
We present a new robust contrastive learning scheme, coined R'enyiCL, which can effectively manage harder augmentations.
Our method is built upon the variational lower bound of R'enyi divergence.
We show that R'enyi contrastive learning objectives perform innate hard negative sampling and easy positive sampling simultaneously.
arXiv Detail & Related papers (2022-08-12T13:37:05Z) - ReSSL: Relational Self-Supervised Learning with Weak Augmentation [68.47096022526927]
Self-supervised learning has achieved great success in learning visual representations without data annotations.
We introduce a novel relational SSL paradigm that learns representations by modeling the relationship between different instances.
Our proposed ReSSL significantly outperforms the previous state-of-the-art algorithms in terms of both performance and training efficiency.
arXiv Detail & Related papers (2021-07-20T06:53:07Z) - Similarity Transfer for Knowledge Distillation [25.042405967561212]
Knowledge distillation is a popular paradigm for learning portable neural networks by transferring the knowledge from a large model into a smaller one.
We propose a novel method called similarity transfer for knowledge distillation (STKD), which aims to fully utilize the similarities between categories of multiple samples.
It shows that STKD substantially has outperformed the vanilla knowledge distillation and has achieved superior accuracy over the state-of-the-art knowledge distillation methods.
arXiv Detail & Related papers (2021-03-18T06:54:59Z) - Doubly Contrastive Deep Clustering [135.7001508427597]
We present a novel Doubly Contrastive Deep Clustering (DCDC) framework, which constructs contrastive loss over both sample and class views.
Specifically, for the sample view, we set the class distribution of the original sample and its augmented version as positive sample pairs.
For the class view, we build the positive and negative pairs from the sample distribution of the class.
In this way, two contrastive losses successfully constrain the clustering results of mini-batch samples in both sample and class level.
arXiv Detail & Related papers (2021-03-09T15:15:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.