Improving the performance of object detection by preserving label
distribution
- URL: http://arxiv.org/abs/2308.14466v1
- Date: Mon, 28 Aug 2023 10:04:06 GMT
- Title: Improving the performance of object detection by preserving label
distribution
- Authors: Heewon Lee, Sangtae Ahn
- Abstract summary: We propose a method that evenly distributes the classes in an image for training and validation.
Our proposed method aims to maintain a uniform class distribution through multi-label stratification.
Our findings indicate that the proposed method can be effectively used on datasets with substantially imbalanced class distribution.
- Score: 1.786745679681672
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object detection is a task that performs position identification and label
classification of objects in images or videos. The information obtained through
this process plays an essential role in various tasks in the field of computer
vision. In object detection, the data utilized for training and validation
typically originate from public datasets that are well-balanced in terms of the
number of objects ascribed to each class in an image. However, in real-world
scenarios, handling datasets with much greater class imbalance, i.e., very
different numbers of objects for each class , is much more common, and this
imbalance may reduce the performance of object detection when predicting unseen
test images. In our study, thus, we propose a method that evenly distributes
the classes in an image for training and validation, solving the class
imbalance problem in object detection. Our proposed method aims to maintain a
uniform class distribution through multi-label stratification. We tested our
proposed method not only on public datasets that typically exhibit balanced
class distribution but also on custom datasets that may have imbalanced class
distribution. We found that our proposed method was more effective on datasets
containing severe imbalance and less data. Our findings indicate that the
proposed method can be effectively used on datasets with substantially
imbalanced class distribution.
Related papers
- DatasetEquity: Are All Samples Created Equal? In The Quest For Equity
Within Datasets [4.833815605196965]
This paper presents a novel method for addressing data imbalance in machine learning.
It computes sample likelihoods based on image appearance using deep perceptual embeddings and clustering.
It then uses these likelihoods to weigh samples differently during training with a proposed $bfGeneralized Focal Loss$ function.
arXiv Detail & Related papers (2023-08-19T02:11:49Z) - Disambiguation of One-Shot Visual Classification Tasks: A Simplex-Based
Approach [8.436437583394998]
We present a strategy which aims at detecting the presence of multiple objects in a given shot.
This strategy is based on identifying the corners of a simplex in a high dimensional space.
We show the ability of the proposed method to slightly, yet statistically significantly, improve accuracy in extreme settings.
arXiv Detail & Related papers (2023-01-16T11:37:05Z) - MUS-CDB: Mixed Uncertainty Sampling with Class Distribution Balancing
for Active Annotation in Aerial Object Detection [40.94800050576902]
Recent aerial object detection models rely on a large amount of labeled training data.
Active learning effectively reduces the data labeling cost by selectively querying the informative and representative unlabelled samples.
We propose a novel active learning method for cost-effective aerial object detection.
arXiv Detail & Related papers (2022-12-06T07:50:00Z) - Contrastive Object Detection Using Knowledge Graph Embeddings [72.17159795485915]
We compare the error statistics of the class embeddings learned from a one-hot approach with semantically structured embeddings from natural language processing or knowledge graphs.
We propose a knowledge-embedded design for keypoint-based and transformer-based object detection architectures.
arXiv Detail & Related papers (2021-12-21T17:10:21Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Sequential Targeting: an incremental learning approach for data
imbalance in text classification [7.455546102930911]
Methods to handle imbalanced datasets are crucial for alleviating distributional skews.
We propose a novel training method, Sequential Targeting(ST), independent of the effectiveness of the representation method.
We demonstrate the effectiveness of our method through experiments on simulated benchmark datasets (IMDB) and data collected from NAVER.
arXiv Detail & Related papers (2020-11-20T04:54:00Z) - A Few-Shot Sequential Approach for Object Counting [63.82757025821265]
We introduce a class attention mechanism that sequentially attends to objects in the image and extracts their relevant features.
The proposed technique is trained on point-level annotations and uses a novel loss function that disentangles class-dependent and class-agnostic aspects of the model.
We present our results on a variety of object-counting/detection datasets, including FSOD and MS COCO.
arXiv Detail & Related papers (2020-07-03T18:23:39Z) - Adaptive Object Detection with Dual Multi-Label Prediction [78.69064917947624]
We propose a novel end-to-end unsupervised deep domain adaptation model for adaptive object detection.
The model exploits multi-label prediction to reveal the object category information in each image.
We introduce a prediction consistency regularization mechanism to assist object detection.
arXiv Detail & Related papers (2020-03-29T04:23:22Z) - Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition
from a Domain Adaptation Perspective [98.70226503904402]
Object frequency in the real world often follows a power law, leading to a mismatch between datasets with long-tailed class distributions.
We propose to augment the classic class-balanced learning by explicitly estimating the differences between the class-conditioned distributions with a meta-learning approach.
arXiv Detail & Related papers (2020-03-24T11:28:42Z) - Pairwise Similarity Knowledge Transfer for Weakly Supervised Object
Localization [53.99850033746663]
We study the problem of learning localization model on target classes with weakly supervised image labels.
In this work, we argue that learning only an objectness function is a weak form of knowledge transfer.
Experiments on the COCO and ILSVRC 2013 detection datasets show that the performance of the localization model improves significantly with the inclusion of pairwise similarity function.
arXiv Detail & Related papers (2020-03-18T17:53:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.