Improving Zero-Shot Models with Label Distribution Priors
- URL: http://arxiv.org/abs/2212.00784v1
- Date: Thu, 1 Dec 2022 18:59:03 GMT
- Title: Improving Zero-Shot Models with Label Distribution Priors
- Authors: Jonathan Kahana, Niv Cohen, Yedid Hoshen
- Abstract summary: We propose a new approach, CLIPPR, which adapts zero-shot models for regression and classification on unlabelled datasets.
We demonstrate an improvement of 28% in mean absolute error on the UTK age regression task.
We also present promising results for classification benchmarks, improving the classification accuracy on the ImageNet dataset by 2.83%, without using any labels.
- Score: 33.51714665243138
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Labeling large image datasets with attributes such as facial age or object
type is tedious and sometimes infeasible. Supervised machine learning methods
provide a highly accurate solution, but require manual labels which are often
unavailable. Zero-shot models (e.g., CLIP) do not require manual labels but are
not as accurate as supervised ones, particularly when the attribute is numeric.
We propose a new approach, CLIPPR (CLIP with Priors), which adapts zero-shot
models for regression and classification on unlabelled datasets. Our method
does not use any annotated images. Instead, we assume a prior over the label
distribution in the dataset. We then train an adapter network on top of CLIP
under two competing objectives: i) minimal change of predictions from the
original CLIP model ii) minimal distance between predicted and prior
distribution of labels. Additionally, we present a novel approach for selecting
prompts for Vision & Language models using a distributional prior. Our method
is effective and presents a significant improvement over the original model. We
demonstrate an improvement of 28% in mean absolute error on the UTK age
regression task. We also present promising results for classification
benchmarks, improving the classification accuracy on the ImageNet dataset by
2.83%, without using any labels.
Related papers
- Enhancing Zero-Shot Vision Models by Label-Free Prompt Distribution Learning and Bias Correcting [55.361337202198925]
Vision-language models, such as CLIP, have shown impressive generalization capacities when using appropriate text descriptions.
We propose a label-Free prompt distribution learning and bias correction framework, dubbed as **Frolic**, which boosts zero-shot performance without the need for labeled data.
arXiv Detail & Related papers (2024-10-25T04:00:45Z) - ReCLIP: Refine Contrastive Language Image Pre-Training with Source Free
Domain Adaptation [20.57370550156505]
ReCLIP is a source-free domain adaptation method for vision-language models.
We demonstrate ReCLIP reduces the average error rate of CLIP from 30.17% to 25.06% on 22 image classification benchmarks.
arXiv Detail & Related papers (2023-08-04T18:11:40Z) - Semi-Supervised Learning with Pseudo-Negative Labels for Image
Classification [14.100569951592417]
We propose a mutual learning framework based on pseudo-negative labels.
By reducing the prediction probability on pseudo-negative labels, the dual model can improve its prediction ability.
Our framework achieves state-of-the-art results on several main benchmarks.
arXiv Detail & Related papers (2023-01-10T14:15:17Z) - NorMatch: Matching Normalizing Flows with Discriminative Classifiers for
Semi-Supervised Learning [8.749830466953584]
Semi-Supervised Learning (SSL) aims to learn a model using a tiny labeled set and massive amounts of unlabeled data.
In this work we introduce a new framework for SSL named NorMatch.
We demonstrate, through numerical and visual results, that NorMatch achieves state-of-the-art performance on several datasets.
arXiv Detail & Related papers (2022-11-17T15:39:18Z) - Conformal Predictor for Improving Zero-shot Text Classification
Efficiency [37.745518881553416]
We reduce the average inference time for NLI- and NSP-based models by 25.6% and 22.2% respectively.
With a suitable CP for each dataset, we reduce the average inference time for NLI- and NSP-based models by 25.6% and 22.2% respectively.
arXiv Detail & Related papers (2022-10-23T05:19:50Z) - Active Learning by Feature Mixing [52.16150629234465]
We propose a novel method for batch active learning called ALFA-Mix.
We identify unlabelled instances with sufficiently-distinct features by seeking inconsistencies in predictions.
We show that inconsistencies in these predictions help discovering features that the model is unable to recognise in the unlabelled instances.
arXiv Detail & Related papers (2022-03-14T12:20:54Z) - Cross-Model Pseudo-Labeling for Semi-Supervised Action Recognition [98.25592165484737]
We propose a more effective pseudo-labeling scheme, called Cross-Model Pseudo-Labeling (CMPL)
CMPL achieves $17.6%$ and $25.1%$ Top-1 accuracy on Kinetics-400 and UCF-101 using only the RGB modality and $1%$ labeled data, respectively.
arXiv Detail & Related papers (2021-12-17T18:59:41Z) - Multi-label Classification with Partial Annotations using Class-aware
Selective Loss [14.3159150577502]
Large-scale multi-label classification datasets are commonly partially annotated.
We analyze the partial labeling problem, then propose a solution based on two key ideas.
With our novel approach, we achieve state-of-the-art results on OpenImages dataset.
arXiv Detail & Related papers (2021-10-21T08:10:55Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - One-bit Supervision for Image Classification [121.87598671087494]
One-bit supervision is a novel setting of learning from incomplete annotations.
We propose a multi-stage training paradigm which incorporates negative label suppression into an off-the-shelf semi-supervised learning algorithm.
arXiv Detail & Related papers (2020-09-14T03:06:23Z) - Automatically Discovering and Learning New Visual Categories with
Ranking Statistics [145.89790963544314]
We tackle the problem of discovering novel classes in an image collection given labelled examples of other classes.
We learn a general-purpose clustering model and use the latter to identify the new classes in the unlabelled data.
We evaluate our approach on standard classification benchmarks and outperform current methods for novel category discovery by a significant margin.
arXiv Detail & Related papers (2020-02-13T18:53:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.