Optimizing Shortfall Risk Metric for Learning Regression Models
        - URL: http://arxiv.org/abs/2505.17777v3
 - Date: Wed, 11 Jun 2025 12:47:35 GMT
 - Title: Optimizing Shortfall Risk Metric for Learning Regression Models
 - Authors: Harish G. Ramaswamy, L. A. Prashanth, 
 - Abstract summary: Empirical risk minimization with a UBSR objective is challenging since UBSR is a non-linear function of the underlying distribution.<n>We devise a bisection-type algorithm, and establish convergence to the UBSR-optimal solution.
 - Score: 0.6445605125467572
 - License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
 - Abstract:   We consider the problem of estimating and optimizing utility-based shortfall risk (UBSR) of a loss, say $(Y - \hat Y)^2$, in the context of a regression problem. Empirical risk minimization with a UBSR objective is challenging since UBSR is a non-linear function of the underlying distribution. We first derive a concentration bound for UBSR estimation using independent and identically distributed (i.i.d.) samples. We then frame the UBSR optimization problem as minimization of a pseudo-linear function in the space of achievable distributions $\mathcal D$ of the loss $(Y- \hat Y)^2$. We construct a gradient oracle for the UBSR objective and a linear minimization oracle (LMO) for the set $\mathcal D$. Using these oracles, we devise a bisection-type algorithm, and establish convergence to the UBSR-optimal solution. 
 
       
      
        Related papers
        - Dynamic Regret Reduces to Kernelized Static Regret [63.36965242404415]
We study dynamic regret in online convex optimization, where the objective is to achieve low cumulative loss relative to an arbitrary benchmark sequence.<n>By constructing a suitable function space in the form of a Reproducing Kernel Hilbert Space (RKHS), our reduction enables us to recover the optimal $R_T(u_1,ldots,u_T) = mathcalO(sqrtsum_t|u_t-u_t-u_t-1|T)$ dynamic regret guarantee.
arXiv  Detail & Related papers  (2025-07-07T21:09:33Z) - A Sample Efficient Alternating Minimization-based Algorithm For Robust   Phase Retrieval [56.67706781191521]
In this work, we present a robust phase retrieval problem where the task is to recover an unknown signal.
Our proposed oracle avoids the need for computationally spectral descent, using a simple gradient step and outliers.
arXiv  Detail & Related papers  (2024-09-07T06:37:23Z) - Alternating Minimization Schemes for Computing   Rate-Distortion-Perception Functions with $f$-Divergence Perception   Constraints [10.564071872770146]
We study the computation of the rate-distortion-perception function (RDPF) for discrete memoryless sources.
We characterize the optimal parametric solutions.
We provide sufficient conditions on the distortion and the perception constraints.
arXiv  Detail & Related papers  (2024-08-27T12:50:12Z) - Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
  Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv  Detail & Related papers  (2024-02-14T07:52:00Z) - Convex Relaxations of ReLU Neural Networks Approximate Global Optima in   Polynomial Time [45.72323731094864]
In this paper, we study the optimality gap between two-layer ReLULU networks regularized with weight decay and their convex relaxations.
Our study sheds new light on understanding why local methods work well.
arXiv  Detail & Related papers  (2024-02-06T01:29:35Z) - Globally Convergent Accelerated Algorithms for Multilinear Sparse
  Logistic Regression with $\ell_0$-constraints [2.323238724742687]
Multilinear logistic regression serves as a powerful tool for the analysis of multidimensional data.
We propose an Accelerated Proximal Alternating Minim-MLSR model to solve the $ell_0$-MLSR.
We also demonstrate that APALM$+$ is globally convergent to a first-order critical point as well as to establish convergence by using the Kurdy-Lojasiewicz property.
arXiv  Detail & Related papers  (2023-09-17T11:05:08Z) - Reinforcement Learning with General Utilities: Simpler Variance
  Reduction and Large State-Action Space [17.366915676628867]
We consider the reinforcement learning problem with general utilities.
Our algorithm achieves $tildemathcalO(epsilon-3)$ and $tildemathcalO(epsilon-2)$ sample complexities.
arXiv  Detail & Related papers  (2023-06-02T18:16:35Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv  Detail & Related papers  (2022-08-11T04:12:50Z) - A framework for bilevel optimization that enables stochastic and global   variance reduction algorithms [21.67411847762289]
Bilevel optimization is a problem of minimizing a value function which involves the arg-minimum of another function.<n>We introduce a novel framework, in which the solution of the inner problem, the solution of the linear system, and the main variable evolve at the same time.<n>We demonstrate that SABA, an adaptation of the celebrated SAGA algorithm in our framework, has $O(frac1T)$ convergence rate.
arXiv  Detail & Related papers  (2022-01-31T18:17:25Z) - Online Estimation and Optimization of Utility-Based Shortfall Risk [0.9999629695552195]
We consider the problem of estimating Utility-Based Shortfall Risk (UBSR)
We cast the UBSR estimation problem as a root finding problem, and propose approximation-based estimations schemes.
We derive non-asymptotic bounds on the estimation error in the number of samples.
arXiv  Detail & Related papers  (2021-11-16T22:16:03Z) - Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free
  Reinforcement Learning [52.76230802067506]
A novel model-free algorithm is proposed to minimize regret in episodic reinforcement learning.
The proposed algorithm employs an em early-settled reference update rule, with the aid of two Q-learning sequences.
The design principle of our early-settled variance reduction method might be of independent interest to other RL settings.
arXiv  Detail & Related papers  (2021-10-09T21:13:48Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv  Detail & Related papers  (2020-07-16T06:44:44Z) - Least Squares Regression with Markovian Data: Fundamental Limits and
  Algorithms [69.45237691598774]
We study the problem of least squares linear regression where the data-points are dependent and are sampled from a Markov chain.
We establish sharp information theoretic minimax lower bounds for this problem in terms of $tau_mathsfmix$.
We propose an algorithm based on experience replay--a popular reinforcement learning technique--that achieves a significantly better error rate.
arXiv  Detail & Related papers  (2020-06-16T04:26:50Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.