Few-shot $\mathbf{1/a}$ Anomalies Feedback : Damage Vision Mining
Opportunity and Embedding Feature Imbalance
- URL: http://arxiv.org/abs/2307.12676v5
- Date: Tue, 29 Aug 2023 14:48:37 GMT
- Title: Few-shot $\mathbf{1/a}$ Anomalies Feedback : Damage Vision Mining
Opportunity and Embedding Feature Imbalance
- Authors: Takato Yasuno
- Abstract summary: imbalanced data problems can be categorised into four types: missing range of target and label valuables, majority-minority class imbalance, foreground background of spatial imbalance, and long-tailed class of pixel-wise imbalance.
In this study, we highlight a one-class anomaly detection application, whether anomalous class or not, and demonstrate clear examples of imbalanced vision datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the past decade, previous balanced datasets have been used to advance
deep learning algorithms for industrial applications. In urban infrastructures
and living environments, damage data mining cannot avoid imbalanced data issues
because of rare unseen events and the high-quality status of improved
operations. For visual inspection, the deteriorated class acquired from the
surface of concrete and steel components are occasionally imbalanced. From
numerous related surveys, we conclude that imbalanced data problems can be
categorised into four types: 1) missing range of target and label valuables, 2)
majority-minority class imbalance, 3) foreground background of spatial
imbalance, and 4) long-tailed class of pixel-wise imbalance. Since 2015, many
imbalanced studies have been conducted using deep-learning approaches,
including regression, image classification, object detection, and semantic
segmentation. However, anomaly detection for imbalanced data is not well known.
In this study, we highlight a one-class anomaly detection application, whether
anomalous class or not, and demonstrate clear examples of imbalanced vision
datasets: medical disease, hazardous behaviour, material deterioration, plant
disease, river sludge, and disaster damage. We provide key results on the
advantage of damage-vision mining, hypothesising that the more effective the
range of the positive ratio, the higher the accuracy gain of the anomalies
feedback. In our imbalanced studies, compared with the balanced case with a
positive ratio of $1/1$, we find that there is an applicable positive ratio
$1/a$ where the accuracy is consistently high. However, the extremely
imbalanced range is from one shot to $1/2a$, the accuracy of which is inferior
to that of the applicable ratio. In contrast, with a positive ratio ranging
over $2/a$, it shifts in the over-mining phase without an effective gain in
accuracy.
Related papers
- Rethinking the Bias of Foundation Model under Long-tailed Distribution [18.80942166783087]
We find the imbalance biases inherited in foundation models on downstream task as parameter imbalance and data imbalance.
During fine-tuning, we observe that parameter imbalance plays a more critical role, while data imbalance can be mitigated using existing re-balancing strategies.
We propose a novel backdoor adjustment method that learns the true causal effect between input samples and labels.
arXiv Detail & Related papers (2025-01-27T11:00:19Z) - Class Imbalance in Object Detection: An Experimental Diagnosis and Study
of Mitigation Strategies [0.5439020425818999]
This study introduces a benchmarking framework utilizing the YOLOv5 single-stage detector to address the problem of foreground-foreground class imbalance.
We scrutinized three established techniques: sampling, loss weighing, and data augmentation.
Our comparative analysis reveals that sampling and loss reweighing methods, while shown to be beneficial in two-stage detector settings, do not translate as effectively in improving YOLOv5's performance.
arXiv Detail & Related papers (2024-03-11T19:06:04Z) - Delving into Semantic Scale Imbalance [45.30062061215943]
We define and quantify the semantic scale of classes, which is used to measure the feature diversity of classes.
We propose semantic-scale-balanced learning, including a general loss improvement scheme and a dynamic re-weighting training framework.
Comprehensive experiments show that dynamic semantic-scale-balanced learning consistently enables the model to perform superiorly on large-scale long-tailed and non-long-tailed natural and medical datasets.
arXiv Detail & Related papers (2022-12-30T09:40:09Z) - Understanding the Impact of Adversarial Robustness on Accuracy Disparity [18.643495650734398]
We decompose the impact of adversarial robustness into two parts: an inherent effect that will degrade the standard accuracy on all classes due to the robustness constraint, and the other caused by the class imbalance ratio.
Our results suggest that the implications may extend to nonlinear models over real-world datasets.
arXiv Detail & Related papers (2022-11-28T20:46:51Z) - Density-Aware Personalized Training for Risk Prediction in Imbalanced
Medical Data [89.79617468457393]
Training models with imbalance rate (class density discrepancy) may lead to suboptimal prediction.
We propose a framework for training models for this imbalance issue.
We demonstrate our model's improved performance in real-world medical datasets.
arXiv Detail & Related papers (2022-07-23T00:39:53Z) - Neural Collapse Inspired Attraction-Repulsion-Balanced Loss for
Imbalanced Learning [97.81549071978789]
We propose Attraction-Repulsion-Balanced Loss (ARB-Loss) to balance the different components of the gradients.
We perform experiments on the large-scale classification and segmentation datasets and our ARB-Loss can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-04-19T08:23:23Z) - Scale-Equivalent Distillation for Semi-Supervised Object Detection [57.59525453301374]
Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals.
We analyze the challenges these methods meet with the empirical experiment results.
We introduce a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance.
arXiv Detail & Related papers (2022-03-23T07:33:37Z) - Balance-Oriented Focal Loss with Linear Scheduling for Anchor Free
Object Detection [1.69146632099647]
We propose Balance-oriented focal loss that can induce balanced learning by considering both background and foreground balance.
By improving the focal loss in terms of balancing foreground classes, our method achieves AP gains of +1.2 in MS-COCO for the anchor free real-time detector.
arXiv Detail & Related papers (2020-12-26T15:24:03Z) - On the Importance of Adaptive Data Collection for Extremely Imbalanced
Pairwise Tasks [94.23884467360521]
We show that state-of-the art models trained on QQP and WikiQA each have only $2.4%$ average precision when evaluated on realistically imbalanced test data.
By creating balanced training data with more informative negative examples, active learning greatly improves average precision to $32.5%$ on QQP and $20.1%$ on WikiQA.
arXiv Detail & Related papers (2020-10-10T21:56:27Z) - Imbalanced Image Classification with Complement Cross Entropy [10.35173901214638]
We study the study of cross entropy which mostly ignores output scores on incorrect classes.
This work discovers that predicted probabilities on incorrect classes improves the prediction accuracy for imbalanced image classification.
The proposed loss makes the ground truth class overwhelm the other classes in terms of softmax probability.
arXiv Detail & Related papers (2020-09-04T13:46:24Z) - Provable tradeoffs in adversarially robust classification [96.48180210364893]
We develop and leverage new tools, including recent breakthroughs from probability theory on robust isoperimetry.
Our results reveal fundamental tradeoffs between standard and robust accuracy that grow when data is imbalanced.
arXiv Detail & Related papers (2020-06-09T09:58:19Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.