Max-Margin Contrastive Learning
- URL: http://arxiv.org/abs/2112.11450v1
- Date: Tue, 21 Dec 2021 18:56:54 GMT
- Title: Max-Margin Contrastive Learning
- Authors: Anshul Shah and Suvrit Sra and Rama Chellappa and Anoop Cherian
- Abstract summary: We present max-margin contrastive learning (MMCL) for unsupervised representation learning.
Our approach selects negatives as the sparse support vectors obtained via a quadratic optimization problem.
We validate our approach on standard vision benchmark datasets, demonstrating better performance in unsupervised representation learning.
- Score: 120.32963353348674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Standard contrastive learning approaches usually require a large number of
negatives for effective unsupervised learning and often exhibit slow
convergence. We suspect this behavior is due to the suboptimal selection of
negatives used for offering contrast to the positives. We counter this
difficulty by taking inspiration from support vector machines (SVMs) to present
max-margin contrastive learning (MMCL). Our approach selects negatives as the
sparse support vectors obtained via a quadratic optimization problem, and
contrastiveness is enforced by maximizing the decision margin. As SVM
optimization can be computationally demanding, especially in an end-to-end
setting, we present simplifications that alleviate the computational burden. We
validate our approach on standard vision benchmark datasets, demonstrating
better performance in unsupervised representation learning over
state-of-the-art, while having better empirical convergence properties.
Related papers
- Provable Optimization for Adversarial Fair Self-supervised Contrastive Learning [49.417414031031264]
This paper studies learning fair encoders in a self-supervised learning setting.
All data are unlabeled and only a small portion of them are annotated with sensitive attributes.
arXiv Detail & Related papers (2024-06-09T08:11:12Z) - Exploiting Pseudo Image Captions for Multimodal Summarization [26.033681302592207]
Cross-modal contrastive learning in vision language pretraining faces the challenge of (partial) false negatives.
We propose a contrastive learning strategy regulated by progressively refined cross-modal similarity, to more accurately optimize MI between an image/text anchor and its negative texts/images.
arXiv Detail & Related papers (2023-05-09T14:47:25Z) - Vision Language Pre-training by Contrastive Learning with Cross-Modal
Similarity Regulation [44.851623239151124]
Cross-modal contrastive learning in vision language pretraining faces the challenge of (partial) false negatives.
We propose a contrastive learning strategy regulated by progressively refined cross-modal similarity, to more accurately optimize MI between an image/text anchor and its negative texts/images.
arXiv Detail & Related papers (2023-05-08T05:53:30Z) - Supervised Contrastive Learning as Multi-Objective Optimization for
Fine-Tuning Large Pre-trained Language Models [3.759936323189417]
Supervised Contrastive Learning (SCL) has been shown to achieve excellent performance in most classification tasks.
In this work, we formulate the SCL problem as a Multi-Objective Optimization problem for the fine-tuning phase of RoBERTa language model.
arXiv Detail & Related papers (2022-09-28T15:13:58Z) - Provable Stochastic Optimization for Global Contrastive Learning: Small
Batch Does Not Harm Performance [53.49803579981569]
We consider a global objective for contrastive learning, which contrasts each positive pair with all negative pairs for an anchor point.
Existing methods such as SimCLR requires a large batch size in order to achieve a satisfactory result.
We propose a memory-efficient optimization algorithm for solving the Global Contrastive Learning of Representations, named SogCLR.
arXiv Detail & Related papers (2022-02-24T22:16:53Z) - MIO : Mutual Information Optimization using Self-Supervised Binary
Contrastive Learning [19.5917119072985]
We model contrastive learning into a binary classification problem to predict if a pair is positive or not.
The proposed method outperforms the state-of-the-art algorithms on benchmark datasets like STL-10, CIFAR-10, CIFAR-100.
arXiv Detail & Related papers (2021-11-24T17:51:29Z) - Solving Inefficiency of Self-supervised Representation Learning [87.30876679780532]
Existing contrastive learning methods suffer from very low learning efficiency.
Under-clustering and over-clustering problems are major obstacles to learning efficiency.
We propose a novel self-supervised learning framework using a median triplet loss.
arXiv Detail & Related papers (2021-04-18T07:47:10Z) - Estimating Average Treatment Effects with Support Vector Machines [77.34726150561087]
Support vector machine (SVM) is one of the most popular classification algorithms in the machine learning literature.
We adapt SVM as a kernel-based weighting procedure that minimizes the maximum mean discrepancy between the treatment and control groups.
We characterize the bias of causal effect estimation arising from this trade-off, connecting the proposed SVM procedure to the existing kernel balancing methods.
arXiv Detail & Related papers (2021-02-23T20:22:56Z) - Fast Rates for Contextual Linear Optimization [52.39202699484225]
We show that a naive plug-in approach achieves regret convergence rates that are significantly faster than methods that directly optimize downstream decision performance.
Our results are overall positive for practice: predictive models are easy and fast to train using existing tools, simple to interpret, and, as we show, lead to decisions that perform very well.
arXiv Detail & Related papers (2020-11-05T18:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.