Your Negative May not Be True Negative: Boosting Image-Text Matching
with False Negative Elimination
- URL: http://arxiv.org/abs/2308.04380v1
- Date: Tue, 8 Aug 2023 16:31:43 GMT
- Title: Your Negative May not Be True Negative: Boosting Image-Text Matching
with False Negative Elimination
- Authors: Haoxuan Li, Yi Bin, Junrong Liao, Yang Yang, Heng Tao Shen
- Abstract summary: We propose a novel False Negative Elimination (FNE) strategy to select negatives via sampling.
The results demonstrate the superiority of our proposed false negative elimination strategy.
- Score: 62.18768931714238
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most existing image-text matching methods adopt triplet loss as the
optimization objective, and choosing a proper negative sample for the triplet
of <anchor, positive, negative> is important for effectively training the
model, e.g., hard negatives make the model learn efficiently and effectively.
However, we observe that existing methods mainly employ the most similar
samples as hard negatives, which may not be true negatives. In other words, the
samples with high similarity but not paired with the anchor may reserve
positive semantic associations, and we call them false negatives. Repelling
these false negatives in triplet loss would mislead the semantic representation
learning and result in inferior retrieval performance. In this paper, we
propose a novel False Negative Elimination (FNE) strategy to select negatives
via sampling, which could alleviate the problem introduced by false negatives.
Specifically, we first construct the distributions of positive and negative
samples separately via their similarities with the anchor, based on the
features extracted from image and text encoders. Then we calculate the false
negative probability of a given sample based on its similarity with the anchor
and the above distributions via the Bayes' rule, which is employed as the
sampling weight during negative sampling process. Since there may not exist any
false negative in a small batch size, we design a memory module with momentum
to retain a large negative buffer and implement our negative sampling strategy
spanning over the buffer. In addition, to make the model focus on hard
negatives, we reassign the sampling weights for the simple negatives with a
cut-down strategy. The extensive experiments are conducted on Flickr30K and
MS-COCO, and the results demonstrate the superiority of our proposed false
negative elimination strategy. The code is available at
https://github.com/LuminosityX/FNE.
Related papers
- Contrastive Learning with Negative Sampling Correction [52.990001829393506]
We propose a novel contrastive learning method named Positive-Unlabeled Contrastive Learning (PUCL)
PUCL treats the generated negative samples as unlabeled samples and uses information from positive samples to correct bias in contrastive loss.
PUCL can be applied to general contrastive learning problems and outperforms state-of-the-art methods on various image and graph classification tasks.
arXiv Detail & Related papers (2024-01-13T11:18:18Z) - SimANS: Simple Ambiguous Negatives Sampling for Dense Text Retrieval [126.22182758461244]
We show that according to the measured relevance scores, the negatives ranked around the positives are generally more informative and less likely to be false negatives.
We propose a simple ambiguous negatives sampling method, SimANS, which incorporates a new sampling probability distribution to sample more ambiguous negatives.
arXiv Detail & Related papers (2022-10-21T07:18:05Z) - Entity Aware Negative Sampling with Auxiliary Loss of False Negative
Prediction for Knowledge Graph Embedding [0.0]
We propose a novel method called Entity Aware Negative Sampling (EANS)
EANS is able to sample negative entities resemble to positive one by adopting Gaussian distribution to the aligned entity index space.
The proposed method can generate high-quality negative samples regardless of negative sample size and effectively mitigate the influence of false negative samples.
arXiv Detail & Related papers (2022-10-12T14:27:51Z) - Negative Sampling for Recommendation [7.758275614033198]
How to effectively sample high-quality negative instances is important for well training a recommendation model.
We argue that a high-quality negative should be both textitinformativeness and textitunbiasedness
arXiv Detail & Related papers (2022-04-02T09:50:19Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Contrastive Attraction and Contrastive Repulsion for Representation
Learning [131.72147978462348]
Contrastive learning (CL) methods learn data representations in a self-supervision manner, where the encoder contrasts each positive sample over multiple negative samples.
Recent CL methods have achieved promising results when pretrained on large-scale datasets, such as ImageNet.
We propose a doubly CL strategy that separately compares positive and negative samples within their own groups, and then proceeds with a contrast between positive and negative groups.
arXiv Detail & Related papers (2021-05-08T17:25:08Z) - Contrastive Learning with Hard Negative Samples [80.12117639845678]
We develop a new family of unsupervised sampling methods for selecting hard negative samples.
A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible.
The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.
arXiv Detail & Related papers (2020-10-09T14:18:53Z) - Simplify and Robustify Negative Sampling for Implicit Collaborative
Filtering [42.832851785261894]
In this paper, we first provide a novel understanding of negative instances by empirically observing that only a few instances are potentially important for model learning.
We then tackle the untouched false negative problem by favouring high-variance samples stored in memory.
Empirical results on two synthetic datasets and three real-world datasets demonstrate both robustness and superiorities of our negative sampling method.
arXiv Detail & Related papers (2020-09-07T19:08:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.