Contrasting quadratic assignments for set-based representation learning
- URL: http://arxiv.org/abs/2205.15814v1
- Date: Tue, 31 May 2022 14:14:36 GMT
- Title: Contrasting quadratic assignments for set-based representation learning
- Authors: Artem Moskalev and Ivan Sosnovik and Volker Fischer and Arnold
Smeulders
- Abstract summary: standard approach to contrastive learning is to maximize the agreement between different views of the data.
In this work, we note that the approach of considering individual pairs cannot account for both intra-set and inter-set similarities.
We propose to go beyond contrasting individual pairs of objects by focusing on contrasting objects as sets.
- Score: 5.142415132534397
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The standard approach to contrastive learning is to maximize the agreement
between different views of the data. The views are ordered in pairs, such that
they are either positive, encoding different views of the same object, or
negative, corresponding to views of different objects. The supervisory signal
comes from maximizing the total similarity over positive pairs, while the
negative pairs are needed to avoid collapse. In this work, we note that the
approach of considering individual pairs cannot account for both intra-set and
inter-set similarities when the sets are formed from the views of the data. It
thus limits the information content of the supervisory signal available to
train representations. We propose to go beyond contrasting individual pairs of
objects by focusing on contrasting objects as sets. For this, we use
combinatorial quadratic assignment theory designed to evaluate set and graph
similarities and derive set-contrastive objective as a regularizer for
contrastive learning methods. We conduct experiments and demonstrate that our
method improves learned representations for the tasks of metric learning and
self-supervised classification.
Related papers
- Rethinking Positive Pairs in Contrastive Learning [19.149235307036324]
We present Hydra, a universal contrastive learning framework for visual representations that extends conventional contrastive learning to accommodate arbitrary pairs.
Our approach is validated using IN1K, where 1K diverse classes compose 500,500 pairs, most of them being distinct.
Our work highlights the value of learning common features of arbitrary pairs and potentially broadens the applicability of contrastive learning techniques on the sample pairs with weak relationships.
arXiv Detail & Related papers (2024-10-23T18:07:18Z) - Discriminative Anchor Learning for Efficient Multi-view Clustering [59.11406089896875]
We propose discriminative anchor learning for multi-view clustering (DALMC)
We learn discriminative view-specific feature representations according to the original dataset.
We build anchors from different views based on these representations, which increase the quality of the shared anchor graph.
arXiv Detail & Related papers (2024-09-25T13:11:17Z) - Entropy Neural Estimation for Graph Contrastive Learning [9.032721248598088]
Contrastive learning on graphs aims at extracting distinguishable high-level representations of nodes.
We propose a simple yet effective subset sampling strategy to contrast pairwise representations between views of a dataset.
We conduct extensive experiments on seven graph benchmarks, and the proposed approach achieves competitive performance.
arXiv Detail & Related papers (2023-07-26T03:55:08Z) - Learning by Sorting: Self-supervised Learning with Group Ordering
Constraints [75.89238437237445]
This paper proposes a new variation of the contrastive learning objective, Group Ordering Constraints (GroCo)
It exploits the idea of sorting the distances of positive and negative pairs and computing the respective loss based on how many positive pairs have a larger distance than the negative pairs, and thus are not ordered correctly.
We evaluate the proposed formulation on various self-supervised learning benchmarks and show that it not only leads to improved results compared to vanilla contrastive learning but also shows competitive performance to comparable methods in linear probing and outperforms current methods in k-NN performance.
arXiv Detail & Related papers (2023-01-05T11:17:55Z) - Non-contrastive representation learning for intervals from well logs [58.70164460091879]
The representation learning problem in the oil & gas industry aims to construct a model that provides a representation based on logging data for a well interval.
One of the possible approaches is self-supervised learning (SSL)
We are the first to introduce non-contrastive SSL for well-logging data.
arXiv Detail & Related papers (2022-09-28T13:27:10Z) - Extending Momentum Contrast with Cross Similarity Consistency
Regularization [5.085461418671174]
We present Extended Momentum Contrast, a self-supervised representation learning method founded upon the legacy of the momentum-encoder unit proposed in the MoCo family configurations.
Under the cross consistency regularization rule, we argue that semantic representations associated with any pair of images (positive or negative) should preserve their cross-similarity.
We report a competitive performance on the standard Imagenet-1K linear head classification benchmark.
arXiv Detail & Related papers (2022-06-07T20:06:56Z) - Self-Supervised Visual Representation Learning with Semantic Grouping [50.14703605659837]
We tackle the problem of learning visual representations from unlabeled scene-centric data.
We propose contrastive learning from data-driven semantic slots, namely SlotCon, for joint semantic grouping and representation learning.
arXiv Detail & Related papers (2022-05-30T17:50:59Z) - ACTIVE:Augmentation-Free Graph Contrastive Learning for Partial
Multi-View Clustering [52.491074276133325]
We propose an augmentation-free graph contrastive learning framework to solve the problem of partial multi-view clustering.
The proposed approach elevates instance-level contrastive learning and missing data inference to the cluster-level, effectively mitigating the impact of individual missing data on clustering.
arXiv Detail & Related papers (2022-03-01T02:32:25Z) - Unsupervised Landmark Learning from Unpaired Data [117.81440795184587]
Recent attempts for unsupervised landmark learning leverage synthesized image pairs that are similar in appearance but different in poses.
We propose a cross-image cycle consistency framework which applies the swapping-reconstruction strategy twice to obtain the final supervision.
Our proposed framework is shown to outperform strong baselines by a large margin.
arXiv Detail & Related papers (2020-06-29T13:57:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.