Balancing Bias and Variance for Active Weakly Supervised Learning
- URL: http://arxiv.org/abs/2206.05682v1
- Date: Sun, 12 Jun 2022 07:15:35 GMT
- Title: Balancing Bias and Variance for Active Weakly Supervised Learning
- Authors: Hitesh Sapkota, Qi Yu
- Abstract summary: Modern multiple instance learning (MIL) models achieve competitive performance at the bag level.
However, instance-level prediction, which is essential for many important applications, is unsatisfactory.
We propose a novel deep subset that aims to boost the instance-level prediction.
Experiments conducted over multiple real-world datasets clearly demonstrate the state-of-the-art instance-level prediction.
- Score: 9.145168943972067
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As a widely used weakly supervised learning scheme, modern multiple instance
learning (MIL) models achieve competitive performance at the bag level.
However, instance-level prediction, which is essential for many important
applications, remains largely unsatisfactory. We propose to conduct novel
active deep multiple instance learning that samples a small subset of
informative instances for annotation, aiming to significantly boost the
instance-level prediction. A variance regularized loss function is designed to
properly balance the bias and variance of instance-level predictions, aiming to
effectively accommodate the highly imbalanced instance distribution in MIL and
other fundamental challenges. Instead of directly minimizing the variance
regularized loss that is non-convex, we optimize a distributionally robust bag
level likelihood as its convex surrogate. The robust bag likelihood provides a
good approximation of the variance based MIL loss with a strong theoretical
guarantee. It also automatically balances bias and variance, making it
effective to identify the potentially positive instances to support active
sampling. The robust bag likelihood can be naturally integrated with a deep
architecture to support deep model training using mini-batches of
positive-negative bag pairs. Finally, a novel P-F sampling function is
developed that combines a probability vector and predicted instance scores,
obtained by optimizing the robust bag likelihood. By leveraging the key MIL
assumption, the sampling function can explore the most challenging bags and
effectively detect their positive instances for annotation, which significantly
improves the instance-level prediction. Experiments conducted over multiple
real-world datasets clearly demonstrate the state-of-the-art instance-level
prediction achieved by the proposed model.
Related papers
- cDP-MIL: Robust Multiple Instance Learning via Cascaded Dirichlet Process [23.266122629592807]
Multiple instance learning (MIL) has been extensively applied to whole slide histoparametric image (WSI) analysis.
The existing aggregation strategy in MIL, which primarily relies on the first-order distance between instances, fails to accurately approximate the true feature distribution of each instance.
We propose a new Bayesian nonparametric framework for multiple instance learning, which adopts a cascade of Dirichlet processes (cDP) to incorporate the instance-to-bag characteristic of the WSIs.
arXiv Detail & Related papers (2024-07-16T07:28:39Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Online progressive instance-balanced sampling for weakly supervised
object detection [0.0]
An online progressive instance-balanced sampling (OPIS) algorithm based on hard sampling and soft sampling is proposed in this paper.
The proposed method can significantly improve the baseline, which is also comparable to many existing state-of-the-art results.
arXiv Detail & Related papers (2022-06-21T12:48:13Z) - Interventional Multi-Instance Learning with Deconfounded Instance-Level
Prediction [29.151629044965983]
We propose a novel interventional multi-instance learning (IMIL) framework to achieve deconfounded instance-level prediction.
Unlike traditional likelihood-based strategies, we design an Expectation-Maximization (EM) algorithm based on causal intervention.
Our IMIL method substantially reduces false positives and outperforms state-of-the-art MIL methods.
arXiv Detail & Related papers (2022-04-20T03:17:36Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - CC-Cert: A Probabilistic Approach to Certify General Robustness of
Neural Networks [58.29502185344086]
In safety-critical machine learning applications, it is crucial to defend models against adversarial attacks.
It is important to provide provable guarantees for deep learning models against semantically meaningful input transformations.
We propose a new universal probabilistic certification approach based on Chernoff-Cramer bounds.
arXiv Detail & Related papers (2021-09-22T12:46:04Z) - Contextual Dropout: An Efficient Sample-Dependent Dropout Module [60.63525456640462]
Dropout has been demonstrated as a simple and effective module to regularize the training process of deep neural networks.
We propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module.
Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.
arXiv Detail & Related papers (2021-03-06T19:30:32Z) - Sparse Network Inversion for Key Instance Detection in Multiple Instance
Learning [24.66638752977373]
Multiple Instance Learning (MIL) involves predicting a single label for a bag of instances, given positive or negative labels at bag-level.
The attention-based deep MIL model is a recent advance in both bag-level classification and key instance detection.
We present a method to improve the attention-based deep MIL model in the task of KID.
arXiv Detail & Related papers (2020-09-07T07:01:59Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - Learning a Unified Sample Weighting Network for Object Detection [113.98404690619982]
Region sampling or weighting is significantly important to the success of modern region-based object detectors.
We argue that sample weighting should be data-dependent and task-dependent.
We propose a unified sample weighting network to predict a sample's task weights.
arXiv Detail & Related papers (2020-06-11T16:19:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.