Exploring the Algorithm-Dependent Generalization of AUPRC Optimization
with List Stability
- URL: http://arxiv.org/abs/2209.13262v1
- Date: Tue, 27 Sep 2022 09:06:37 GMT
- Title: Exploring the Algorithm-Dependent Generalization of AUPRC Optimization
with List Stability
- Authors: Peisong Wen, Qianqian Xu, Zhiyong Yang, Yuan He, Qingming Huang
- Abstract summary: optimization of the Area Under the Precision-Recall Curve (AUPRC) is a crucial problem for machine learning.
In this work, we present the first trial in the single-dependent generalization of AUPRC optimization.
Experiments on three image retrieval datasets on speak to the effectiveness and soundness of our framework.
- Score: 107.65337427333064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic optimization of the Area Under the Precision-Recall Curve (AUPRC)
is a crucial problem for machine learning. Although various algorithms have
been extensively studied for AUPRC optimization, the generalization is only
guaranteed in the multi-query case. In this work, we present the first trial in
the single-query generalization of stochastic AUPRC optimization. For sharper
generalization bounds, we focus on algorithm-dependent generalization. There
are both algorithmic and theoretical obstacles to our destination. From an
algorithmic perspective, we notice that the majority of existing stochastic
estimators are biased only when the sampling strategy is biased, and is
leave-one-out unstable due to the non-decomposability. To address these issues,
we propose a sampling-rate-invariant unbiased stochastic estimator with
superior stability. On top of this, the AUPRC optimization is formulated as a
composition optimization problem, and a stochastic algorithm is proposed to
solve this problem. From a theoretical perspective, standard techniques of the
algorithm-dependent generalization analysis cannot be directly applied to such
a listwise compositional optimization problem. To fill this gap, we extend the
model stability from instancewise losses to listwise losses and bridge the
corresponding generalization and stability. Additionally, we construct state
transition matrices to describe the recurrence of the stability, and simplify
calculations by matrix spectrum. Practically, experimental results on three
image retrieval datasets on speak to the effectiveness and soundness of our
framework.
Related papers
- Stability and Generalization for Stochastic Recursive Momentum-based Algorithms for (Strongly-)Convex One to $K$-Level Stochastic Optimizations [20.809499420384256]
STORM-based algorithms have been widely developed to solve one to $K$-level ($K geq 3$) optimization problems.
This paper provides a comprehensive analysis of three representative STORM-based algorithms.
arXiv Detail & Related papers (2024-07-07T07:07:04Z) - Quantization-Based Optimization: Alternative Stochastic Approximation of
Global Optimization [0.0]
We propose a global optimization algorithm based on quantizing the energy level of an objective function in an NP-hard problem.
Numerical experiments show that the proposed algorithm outperforms conventional learning methods in solving NP-hard optimization problems.
arXiv Detail & Related papers (2022-11-08T03:01:45Z) - Stochastic Mirror Descent for Large-Scale Sparse Recovery [13.500750042707407]
We discuss an application of quadratic Approximation to statistical estimation of high-dimensional sparse parameters.
We show that the proposed algorithm attains the optimal convergence of the estimation error under weak assumptions on the regressor distribution.
arXiv Detail & Related papers (2022-10-23T23:23:23Z) - Iterative regularization for low complexity regularizers [18.87017835436693]
Iterative regularization exploits the implicit bias of an optimization algorithm to regularize ill-posed problems.
We propose and study the first iterative regularization procedure able to handle biases described by non smooth and non strongly convex functionals.
arXiv Detail & Related papers (2022-02-01T14:09:00Z) - Outlier-Robust Sparse Estimation via Non-Convex Optimization [73.18654719887205]
We explore the connection between high-dimensional statistics and non-robust optimization in the presence of sparsity constraints.
We develop novel and simple optimization formulations for these problems.
As a corollary, we obtain that any first-order method that efficiently converges to station yields an efficient algorithm for these tasks.
arXiv Detail & Related papers (2021-09-23T17:38:24Z) - Stability and Generalization for Randomized Coordinate Descent [19.687456295228156]
There is no work studying how the models trained by RCD would generalize to test examples.
In this paper, we initialize the generalization analysis of RCD by leveraging the powerful tool of algorithmic stability.
Our analysis shows that RCD enjoys better stability as compared to gradient descent.
arXiv Detail & Related papers (2021-08-17T02:52:50Z) - Stochastic Optimization of Areas Under Precision-Recall Curves with
Provable Convergence [66.83161885378192]
Area under ROC (AUROC) and precision-recall curves (AUPRC) are common metrics for evaluating classification performance for imbalanced problems.
We propose a technical method to optimize AUPRC for deep learning.
arXiv Detail & Related papers (2021-04-18T06:22:21Z) - Recent Theoretical Advances in Non-Convex Optimization [56.88981258425256]
Motivated by recent increased interest in analysis of optimization algorithms for non- optimization in deep networks and other problems in data, we give an overview of recent results of theoretical optimization algorithms for non- optimization.
arXiv Detail & Related papers (2020-12-11T08:28:51Z) - A Dynamical Systems Approach for Convergence of the Bayesian EM
Algorithm [59.99439951055238]
We show how (discrete-time) Lyapunov stability theory can serve as a powerful tool to aid, or even lead, in the analysis (and potential design) of optimization algorithms that are not necessarily gradient-based.
The particular ML problem that this paper focuses on is that of parameter estimation in an incomplete-data Bayesian framework via the popular optimization algorithm known as maximum a posteriori expectation-maximization (MAP-EM)
We show that fast convergence (linear or quadratic) is achieved, which could have been difficult to unveil without our adopted S&C approach.
arXiv Detail & Related papers (2020-06-23T01:34:18Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.