Sample selection for efficient image annotation
- URL: http://arxiv.org/abs/2105.04678v1
- Date: Mon, 10 May 2021 21:25:10 GMT
- Title: Sample selection for efficient image annotation
- Authors: Bishwo Adhikari, Esa Rahtu, Heikki Huttunen
- Abstract summary: Supervised object detection has been proven to be successful in many benchmark datasets achieving human-level performances.
We propose an efficient image selection approach that samples the most informative images from the unlabeled dataset.
Our method can reduce up to 80% of manual annotation workload, compared to full manual labeling setting, and performs better than random sampling.
- Score: 14.695979686066066
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Supervised object detection has been proven to be successful in many
benchmark datasets achieving human-level performances. However, acquiring a
large amount of labeled image samples for supervised detection training is
tedious, time-consuming, and costly. In this paper, we propose an efficient
image selection approach that samples the most informative images from the
unlabeled dataset and utilizes human-machine collaboration in an iterative
train-annotate loop. Image features are extracted by the CNN network followed
by the similarity score calculation, Euclidean distance. Unlabeled images are
then sampled into different approaches based on the similarity score. The
proposed approach is straightforward, simple and sampling takes place prior to
the network training. Experiments on datasets show that our method can reduce
up to 80% of manual annotation workload, compared to full manual labeling
setting, and performs better than random sampling.
Related papers
- Pre-Trained Vision-Language Models as Partial Annotators [40.89255396643592]
Pre-trained vision-language models learn massive data to model unified representations of images and natural languages.
In this paper, we investigate a novel "pre-trained annotating - weakly-supervised learning" paradigm for pre-trained model application and experiment on image classification tasks.
arXiv Detail & Related papers (2024-05-23T17:17:27Z) - DREAM+: Efficient Dataset Distillation by Bidirectional Representative
Matching [40.18223537419178]
We propose a novel dataset matching strategy called DREAM+, which selects representative original images for bidirectional matching.
DREAM+ significantly reduces the number of distillation iterations by more than 15 times without affecting performance.
Given sufficient training time, DREAM+ can further improve the performance and achieve state-of-the-art results.
arXiv Detail & Related papers (2023-10-23T15:55:30Z) - A Semi-Paired Approach For Label-to-Image Translation [6.888253564585197]
We introduce the first semi-supervised (semi-paired) framework for label-to-image translation.
In the semi-paired setting, the model has access to a small set of paired data and a larger set of unpaired images and labels.
We propose a training algorithm for this shared network, and we present a rare classes sampling algorithm to focus on under-represented classes.
arXiv Detail & Related papers (2023-06-23T16:13:43Z) - Semi-Supervised Image Captioning by Adversarially Propagating Labeled
Data [95.0476489266988]
We present a novel data-efficient semi-supervised framework to improve the generalization of image captioning models.
Our proposed method trains a captioner to learn from a paired data and to progressively associate unpaired data.
Our extensive and comprehensive empirical results both on (1) image-based and (2) dense region-based captioning datasets followed by comprehensive analysis on the scarcely-paired dataset.
arXiv Detail & Related papers (2023-01-26T15:25:43Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - CropMix: Sampling a Rich Input Distribution via Multi-Scale Cropping [97.05377757299672]
We present a simple method, CropMix, for producing a rich input distribution from the original dataset distribution.
CropMix can be seamlessly applied to virtually any training recipe and neural network architecture performing classification tasks.
We show that CropMix is of benefit to both contrastive learning and masked image modeling towards more powerful representations.
arXiv Detail & Related papers (2022-05-31T16:57:28Z) - Active Learning for Deep Visual Tracking [51.5063680734122]
Convolutional neural networks (CNNs) have been successfully applied to the single target tracking task in recent years.
In this paper, we propose an active learning method for deep visual tracking, which selects and annotates the unlabeled samples to train the deep CNNs model.
Under the guidance of active learning, the tracker based on the trained deep CNNs model can achieve competitive tracking performance while reducing the labeling cost.
arXiv Detail & Related papers (2021-10-17T11:47:56Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - Transform consistency for learning with noisy labels [9.029861710944704]
We propose a method to identify clean samples only using one single network.
Clean samples prefer to reach consistent predictions for the original images and the transformed images.
In order to mitigate the negative influence of noisy labels, we design a classification loss by using the off-line hard labels and on-line soft labels.
arXiv Detail & Related papers (2021-03-25T14:33:13Z) - CSI: Novelty Detection via Contrastive Learning on Distributionally
Shifted Instances [77.28192419848901]
We propose a simple, yet effective method named contrasting shifted instances (CSI)
In addition to contrasting a given sample with other instances as in conventional contrastive learning methods, our training scheme contrasts the sample with distributionally-shifted augmentations of itself.
Our experiments demonstrate the superiority of our method under various novelty detection scenarios.
arXiv Detail & Related papers (2020-07-16T08:32:56Z) - Enhancing Few-Shot Image Classification with Unlabelled Examples [18.03136114355549]
We develop a transductive meta-learning method that uses unlabelled instances to improve few-shot image classification performance.
Our approach combines a regularized neural adaptive feature extractor to achieve improved test-time classification accuracy using unlabelled data.
arXiv Detail & Related papers (2020-06-17T05:42:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.