Provable Multi-instance Deep AUC Maximization with Stochastic Pooling
- URL: http://arxiv.org/abs/2305.08040v4
- Date: Tue, 6 Jun 2023 05:32:49 GMT
- Title: Provable Multi-instance Deep AUC Maximization with Stochastic Pooling
- Authors: Dixian Zhu, Bokun Wang, Zhi Chen, Yaxing Wang, Milan Sonka, Xiaodong
Wu, Tianbao Yang
- Abstract summary: This paper considers a novel application of deep AUC (DAM) for multi-instance learning (MIL)
A single class label is assigned to a bag of instances (e.g., multiple 2D slices of a scan for a patient)
- Score: 39.46116380220933
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper considers a novel application of deep AUC maximization (DAM) for
multi-instance learning (MIL), in which a single class label is assigned to a
bag of instances (e.g., multiple 2D slices of a CT scan for a patient). We
address a neglected yet non-negligible computational challenge of MIL in the
context of DAM, i.e., bag size is too large to be loaded into {GPU} memory for
backpropagation, which is required by the standard pooling methods of MIL. To
tackle this challenge, we propose variance-reduced stochastic pooling methods
in the spirit of stochastic optimization by formulating the loss function over
the pooled prediction as a multi-level compositional function. By synthesizing
techniques from stochastic compositional optimization and non-convex min-max
optimization, we propose a unified and provable muli-instance DAM (MIDAM)
algorithm with stochastic smoothed-max pooling or stochastic attention-based
pooling, which only samples a few instances for each bag to compute a
stochastic gradient estimator and to update the model parameter. We establish a
similar convergence rate of the proposed MIDAM algorithm as the
state-of-the-art DAM algorithms. Our extensive experiments on conventional MIL
datasets and medical datasets demonstrate the superiority of our MIDAM
algorithm.
Related papers
- Diffusion Stochastic Optimization for Min-Max Problems [33.73046548872663]
The optimistic gradient method is useful in addressing minimax optimization problems.
Motivated by the observation that the conventional version suffers from the need for a large batch size, we introduce and analyze a new formulation termed Samevareps-generativeOGOG.
arXiv Detail & Related papers (2024-01-26T01:16:59Z) - Federated Conditional Stochastic Optimization [110.513884892319]
Conditional optimization has found in a wide range of machine learning tasks, such as in-variant learning tasks, AUPRC, andAML.
This paper proposes algorithms for distributed federated learning.
arXiv Detail & Related papers (2023-10-04T01:47:37Z) - Algorithme EM r\'egularis\'e [0.0]
This paper presents a regularized version of the EM algorithm that efficiently uses prior knowledge to cope with a small sample size.
Experiments on real data highlight the good performance of the proposed algorithm for clustering purposes.
arXiv Detail & Related papers (2023-07-04T23:19:25Z) - Stochastic First-Order Learning for Large-Scale Flexibly Tied Gaussian
Mixture Model [3.4546761246181696]
We propose a new optimization algorithm on the manifold of Gaussian Mixture Models (GMMs)
We observe that methods can outperform the expectation-maximization algorithm in terms of attaining better likelihood, needing fewer epochs for convergence, and consuming less time per each epoch.
arXiv Detail & Related papers (2022-12-11T04:24:52Z) - Robust Multi-view Registration of Point Sets with Laplacian Mixture
Model [25.865100974015412]
We propose a novel probabilistic generative method to align multiple point sets based on the heavy-tailed Laplacian distribution.
We demonstrate the advantages of our method by comparing it with representative state-of-the-art approaches on benchmark challenging data sets.
arXiv Detail & Related papers (2021-10-26T14:49:09Z) - On Stochastic Moving-Average Estimators for Non-Convex Optimization [105.22760323075008]
In this paper, we demonstrate the power of a widely used estimator based on moving average (SEMA) problems.
For all these-the-art results, we also present the results for all these-the-art problems.
arXiv Detail & Related papers (2021-04-30T08:50:24Z) - Parallel Stochastic Mirror Descent for MDPs [72.75921150912556]
We consider the problem of learning the optimal policy for infinite-horizon Markov decision processes (MDPs)
Some variant of Mirror Descent is proposed for convex programming problems with Lipschitz-continuous functionals.
We analyze this algorithm in a general case and obtain an estimate of the convergence rate that does not accumulate errors during the operation of the method.
arXiv Detail & Related papers (2021-02-27T19:28:39Z) - Stochastic Hard Thresholding Algorithms for AUC Maximization [49.00683387735522]
We develop a hard thresholding algorithm for AUC in distributiond classification.
We conduct experiments to show the efficiency and effectiveness of the proposed algorithms.
arXiv Detail & Related papers (2020-11-04T16:49:29Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - An Adaptive EM Accelerator for Unsupervised Learning of Gaussian Mixture
Models [0.7340845393655052]
We propose an Anderson Acceleration scheme for the adaptive Expectation-Maximization (EM) algorithm for unsupervised learning.
The proposed algorithm is able to determine the optimal number of mixture components autonomously, and converges to the optimal solution much faster than its non-accelerated version.
arXiv Detail & Related papers (2020-09-26T22:55:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.