Adaptive Self-Training for Object Detection
- URL: http://arxiv.org/abs/2212.05911v2
- Date: Thu, 23 Nov 2023 16:49:58 GMT
- Title: Adaptive Self-Training for Object Detection
- Authors: Renaud Vandeghen and Gilles Louppe and Marc Van Droogenbroeck
- Abstract summary: We introduce our method Self-Training for Object Detection (ASTOD)
ASTOD determines without cost a threshold value based directly on the ground value of the score histogram.
We use different views of the unlabeled images during the pseudo-labeling step to reduce the number of missed predictions.
- Score: 13.07105239116411
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has emerged as an effective solution for solving the task of
object detection in images but at the cost of requiring large labeled datasets.
To mitigate this cost, semi-supervised object detection methods, which consist
in leveraging abundant unlabeled data, have been proposed and have already
shown impressive results. However, most of these methods require linking a
pseudo-label to a ground-truth object by thresholding. In previous works, this
threshold value is usually determined empirically, which is time consuming, and
only done for a single data distribution. When the domain, and thus the data
distribution, changes, a new and costly parameter search is necessary. In this
work, we introduce our method Adaptive Self-Training for Object Detection
(ASTOD), which is a simple yet effective teacher-student method. ASTOD
determines without cost a threshold value based directly on the ground value of
the score histogram. To improve the quality of the teacher predictions, we also
propose a novel pseudo-labeling procedure. We use different views of the
unlabeled images during the pseudo-labeling step to reduce the number of missed
predictions and thus obtain better candidate labels. Our teacher and our
student are trained separately, and our method can be used in an iterative
fashion by replacing the teacher by the student. On the MS-COCO dataset, our
method consistently performs favorably against state-of-the-art methods that do
not require a threshold parameter, and shows competitive results with methods
that require a parameter sweep search. Additional experiments with respect to a
supervised baseline on the DIOR dataset containing satellite images lead to
similar conclusions, and prove that it is possible to adapt the score threshold
automatically in self-training, regardless of the data distribution. The code
is available at https:// github.com/rvandeghen/ASTOD
Related papers
- Deep Active Learning with Manifold-preserving Trajectory Sampling [2.0717982775472206]
Active learning (AL) is for optimizing the selection of unlabeled data for annotation (labeling)
Existing deep AL methods arguably suffer from bias incurred by clabeled data, which takes a much lower percentage than unlabeled data in AL context.
We propose a novel method, namely Manifold-Preserving Trajectory Sampling (MPTS), aiming to enforce the feature space learned from labeled data to represent a more accurate manifold.
arXiv Detail & Related papers (2024-10-21T03:04:09Z) - DST-Det: Simple Dynamic Self-Training for Open-Vocabulary Object Detection [72.25697820290502]
This work introduces a straightforward and efficient strategy to identify potential novel classes through zero-shot classification.
We refer to this approach as the self-training strategy, which enhances recall and accuracy for novel classes without requiring extra annotations, datasets, and re-training.
Empirical evaluations on three datasets, including LVIS, V3Det, and COCO, demonstrate significant improvements over the baseline performance.
arXiv Detail & Related papers (2023-10-02T17:52:24Z) - Non-iterative optimization of pseudo-labeling thresholds for training
object detection models from multiple datasets [2.1485350418225244]
We propose a non-iterative method to optimize pseudo-labeling thresholds for learning object detection from a collection of low-cost datasets.
We experimentally demonstrate that our proposed method achieves an mAP comparable to that of grid search on the COCO and VOC datasets.
arXiv Detail & Related papers (2022-10-19T00:31:34Z) - Weighted Distillation with Unlabeled Examples [15.825078347452024]
Distillation with unlabeled examples is a popular and powerful method for training deep neural networks in settings where the amount of labeled data is limited.
This paper proposes a principled approach for addressing this issue based on a ''debiasing'' reweighting of the student's loss function tailored to the distillation training paradigm.
arXiv Detail & Related papers (2022-10-13T04:08:56Z) - Improving Localization for Semi-Supervised Object Detection [3.5493798890908104]
We introduce an additional classification task for bounding box localization to improve the filtering of predicted bounding boxes.
Our experiments show that our IL-net increases SSOD performance by 1.14% AP on dataset in limited-annotation regime.
arXiv Detail & Related papers (2022-06-21T08:39:38Z) - Label Matching Semi-Supervised Object Detection [85.99282969977541]
Semi-supervised object detection has made significant progress with the development of mean teacher driven self-training.
Label mismatch problem is not yet fully explored in the previous works, leading to severe confirmation bias during self-training.
We propose a simple yet effective LabelMatch framework from two different yet complementary perspectives.
arXiv Detail & Related papers (2022-06-14T05:59:41Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - SLADE: A Self-Training Framework For Distance Metric Learning [75.54078592084217]
We present a self-training framework, SLADE, to improve retrieval performance by leveraging additional unlabeled data.
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
We then train a student model on both labels and pseudo labels to generate final feature embeddings.
arXiv Detail & Related papers (2020-11-20T08:26:10Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z) - Deep Active Learning for Biased Datasets via Fisher Kernel
Self-Supervision [5.352699766206807]
Active learning (AL) aims to minimize labeling efforts for data-demanding deep neural networks (DNNs)
We propose a low-complexity method for feature density matching using self-supervised Fisher kernel (FK)
Our method outperforms state-of-the-art methods on MNIST, SVHN, and ImageNet classification while requiring only 1/10th of processing.
arXiv Detail & Related papers (2020-03-01T03:56:32Z) - Learning with Out-of-Distribution Data for Audio Classification [60.48251022280506]
We show that detecting and relabelling certain OOD instances, rather than discarding them, can have a positive effect on learning.
The proposed method is shown to improve the performance of convolutional neural networks by a significant margin.
arXiv Detail & Related papers (2020-02-11T21:08:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.