Trip-ROMA: Self-Supervised Learning with Triplets and Random Mappings
- URL: http://arxiv.org/abs/2107.10419v3
- Date: Thu, 24 Aug 2023 03:09:41 GMT
- Title: Trip-ROMA: Self-Supervised Learning with Triplets and Random Mappings
- Authors: Wenbin Li, Xuesong Yang, Meihao Kong, Lei Wang, Jing Huo, Yang Gao and
Jiebo Luo
- Abstract summary: We show that a simple Triplet-based loss can achieve surprisingly good performance without requiring large batches or asymmetry designs.
To alleviate the over-fitting problem in small data regimes, we propose a simple plug-and-play RandOm MApping (ROMA) strategy.
- Score: 59.32440962369532
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive self-supervised learning (SSL) methods, such as MoCo and SimCLR,
have achieved great success in unsupervised visual representation learning.
They rely on a large number of negative pairs and thus require either large
memory banks or large batches. Some recent non-contrastive SSL methods, such as
BYOL and SimSiam, attempt to discard negative pairs and have also shown
remarkable performance. To avoid collapsed solutions caused by not using
negative pairs, these methods require non-trivial asymmetry designs. However,
in small data regimes, we can not obtain a sufficient number of negative pairs
or effectively avoid the over-fitting problem when negatives are not used at
all. To address this situation, we argue that negative pairs are still
important but one is generally sufficient for each positive pair. We show that
a simple Triplet-based loss (Trip) can achieve surprisingly good performance
without requiring large batches or asymmetry designs. Moreover, to alleviate
the over-fitting problem in small data regimes and further enhance the effect
of Trip, we propose a simple plug-and-play RandOm MApping (ROMA) strategy by
randomly mapping samples into other spaces and requiring these randomly
projected samples to satisfy the same relationship indicated by the triplets.
Integrating the triplet-based loss with random mapping, we obtain the proposed
method Trip-ROMA. Extensive experiments, including unsupervised representation
learning and unsupervised few-shot learning, have been conducted on ImageNet-1K
and seven small datasets. They successfully demonstrate the effectiveness of
Trip-ROMA and consistently show that ROMA can further effectively boost other
SSL methods. Code is available at https://github.com/WenbinLee/Trip-ROMA.
Related papers
- Decoupled Contrastive Multi-View Clustering with High-Order Random Walks [25.03805821839733]
We propose a novel robust method dubbed decoupled contrastive multi-view clustering with high-order random walks (DIVIDE)
In brief, DIVIDE leverages random walks to progressively identify data pairs in a global instead of local manner.
DIVIDE could identify in-neighborhood negatives and out-of-neighborhood positives.
arXiv Detail & Related papers (2023-08-22T03:45:13Z) - Self-Supervised Monocular Depth Estimation: Solving the Edge-Fattening
Problem [39.82550656611876]
Triplet loss, popular for metric learning, has made a great success in many computer vision tasks.
We show two drawbacks of the raw triplet loss in MDE and demonstrate our problem-driven redesigns.
arXiv Detail & Related papers (2022-10-02T03:08:59Z) - Non-contrastive representation learning for intervals from well logs [58.70164460091879]
The representation learning problem in the oil & gas industry aims to construct a model that provides a representation based on logging data for a well interval.
One of the possible approaches is self-supervised learning (SSL)
We are the first to introduce non-contrastive SSL for well-logging data.
arXiv Detail & Related papers (2022-09-28T13:27:10Z) - Modality-Aware Triplet Hard Mining for Zero-shot Sketch-Based Image
Retrieval [51.42470171051007]
This paper tackles the Zero-Shot Sketch-Based Image Retrieval (ZS-SBIR) problem from the viewpoint of cross-modality metric learning.
By combining two fundamental learning approaches in DML, e.g., classification training and pairwise training, we set up a strong baseline for ZS-SBIR.
We show that Modality-Aware Triplet Hard Mining (MATHM) enhances the baseline with three types of pairwise learning.
arXiv Detail & Related papers (2021-12-15T08:36:44Z) - LoOp: Looking for Optimal Hard Negative Embeddings for Deep Metric
Learning [17.571160136568455]
We propose a novel approach that looks for optimal hard negatives (LoOp) in the embedding space.
Unlike mining-based methods, our approach considers the entire space between pairs of embeddings to calculate the optimal hard negatives.
arXiv Detail & Related papers (2021-08-20T19:21:33Z) - Semi-Supervised Metric Learning: A Deep Resurrection [22.918651280720855]
Semi-Supervised DML (SSDML) tries to learn a metric using a few labeled examples, and abundantly available unlabeled examples.
We propose a graph-based approach that first propagates the affinities between the pairs of examples.
We impose Metricity constraint on the metric parameters, as it leads to a better performance.
arXiv Detail & Related papers (2021-05-10T12:28:45Z) - Rethinking Deep Contrastive Learning with Embedding Memory [58.66613563148031]
Pair-wise loss functions have been extensively studied and shown to continuously improve the performance of deep metric learning (DML)
We provide a new methodology for systematically studying weighting strategies of various pair-wise loss functions, and rethink pair weighting with an embedding memory.
arXiv Detail & Related papers (2021-03-25T17:39:34Z) - Contrastive Learning with Hard Negative Samples [80.12117639845678]
We develop a new family of unsupervised sampling methods for selecting hard negative samples.
A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible.
The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.
arXiv Detail & Related papers (2020-10-09T14:18:53Z) - Whitening for Self-Supervised Representation Learning [129.57407186848917]
We propose a new loss function for self-supervised representation learning (SSL) based on the whitening of latent-space features.
Our solution does not require asymmetric networks and it is conceptually simple.
arXiv Detail & Related papers (2020-07-13T12:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.