Empirical Risk Minimization with Relative Entropy Regularization:
Optimality and Sensitivity Analysis
- URL: http://arxiv.org/abs/2202.04385v1
- Date: Wed, 9 Feb 2022 10:55:14 GMT
- Title: Empirical Risk Minimization with Relative Entropy Regularization:
Optimality and Sensitivity Analysis
- Authors: Samir M. Perlaza and Gaetan Bisson and I\~naki Esnaola and Alain
Jean-Marie and Stefano Rini
- Abstract summary: The sensitivity of the expected empirical risk to deviations from the solution of the ERM-RER problem is studied.
The expectation of the sensitivity is upper bounded, up to a constant factor, by the square root of the lautum information between the models and the datasets.
- Score: 7.953455469099826
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The optimality and sensitivity of the empirical risk minimization problem
with relative entropy regularization (ERM-RER) are investigated for the case in
which the reference is a sigma-finite measure instead of a probability measure.
This generalization allows for a larger degree of flexibility in the
incorporation of prior knowledge over the set of models. In this setting, the
interplay of the regularization parameter, the reference measure, the risk
function, and the empirical risk induced by the solution of the ERM-RER problem
is characterized. This characterization yields necessary and sufficient
conditions for the existence of a regularization parameter that achieves an
arbitrarily small empirical risk with arbitrarily high probability. The
sensitivity of the expected empirical risk to deviations from the solution of
the ERM-RER problem is studied. The sensitivity is then used to provide upper
and lower bounds on the expected empirical risk. Moreover, it is shown that the
expectation of the sensitivity is upper bounded, up to a constant factor, by
the square root of the lautum information between the models and the datasets.
Related papers
- Error Bounds of Supervised Classification from Information-Theoretic Perspective [0.0]
We explore bounds on the expected risk when using deep neural networks for supervised classification from an information theoretic perspective.
We introduce model risk and fitting error, which are derived from further decomposing the empirical risk.
arXiv Detail & Related papers (2024-06-07T01:07:35Z) - Continuous-time Risk-sensitive Reinforcement Learning via Quadratic Variation Penalty [5.710971447109951]
This paper studies continuous-time risk-sensitive reinforcement learning (RL)
I highlight that the conventional policy gradient representation is inadequate for risk-sensitive problems due to the nonlinear nature of quadratic variation.
I prove the convergence of the proposed algorithm for Merton's investment problem and quantify the impact of temperature parameter on the behavior of the learning procedure.
arXiv Detail & Related papers (2024-04-19T03:05:41Z) - Provable Risk-Sensitive Distributional Reinforcement Learning with
General Function Approximation [54.61816424792866]
We introduce a general framework on Risk-Sensitive Distributional Reinforcement Learning (RS-DisRL), with static Lipschitz Risk Measures (LRM) and general function approximation.
We design two innovative meta-algorithms: textttRS-DisRL-M, a model-based strategy for model-based function approximation, and textttRS-DisRL-V, a model-free approach for general value function approximation.
arXiv Detail & Related papers (2024-02-28T08:43:18Z) - Model-Based Epistemic Variance of Values for Risk-Aware Policy Optimization [59.758009422067]
We consider the problem of quantifying uncertainty over expected cumulative rewards in model-based reinforcement learning.
We propose a new uncertainty Bellman equation (UBE) whose solution converges to the true posterior variance over values.
We introduce a general-purpose policy optimization algorithm, Q-Uncertainty Soft Actor-Critic (QU-SAC) that can be applied for either risk-seeking or risk-averse policy optimization.
arXiv Detail & Related papers (2023-12-07T15:55:58Z) - Out-of-Distribution Optimality of Invariant Risk Minimization [20.389816785165333]
Invariant Risk Minimization (IRM) is considered to be a promising approach to minimize the o.o.d. risk.
This paper rigorously proves that a solution to the bi-level optimization problem minimizes the o.o.d. risk under certain conditions.
arXiv Detail & Related papers (2023-07-22T03:31:15Z) - Analysis of the Relative Entropy Asymmetry in the Regularization of
Empirical Risk Minimization [70.540936204654]
The effect of the relative entropy asymmetry is analyzed in the empirical risk minimization with relative entropy regularization (ERM-RER) problem.
A novel regularization is introduced, coined Type-II regularization, that allows for solutions to the ERM-RER problem with a support that extends outside the support of the reference measure.
arXiv Detail & Related papers (2023-06-12T13:56:28Z) - On the Variance, Admissibility, and Stability of Empirical Risk
Minimization [80.26309576810844]
Empirical Risk Minimization (ERM) with squared loss may attain minimax suboptimal error rates.
We show that under mild assumptions, the suboptimality of ERM must be due to large bias rather than variance.
We also show that our estimates imply stability of ERM, complementing the main result of Caponnetto and Rakhlin (2006) for non-Donsker classes.
arXiv Detail & Related papers (2023-05-29T15:25:48Z) - Empirical Risk Minimization with Relative Entropy Regularization [6.815730801645783]
The empirical risk minimization (ERM) problem with relative entropy regularization (ERM-RER) is investigated.
The solution to this problem, if it exists, is shown to be a unique probability measure, mutually absolutely continuous with the reference measure.
For a fixed dataset and under a specific condition, the empirical risk is shown to be a sub-Gaussian random variable.
arXiv Detail & Related papers (2022-11-12T09:41:02Z) - The Risks of Invariant Risk Minimization [52.7137956951533]
Invariant Risk Minimization is an objective based on the idea for learning deep, invariant features of data.
We present the first analysis of classification under the IRM objective--as well as these recently proposed alternatives--under a fairly natural and general model.
We show that IRM can fail catastrophically unless the test data are sufficiently similar to the training distribution--this is precisely the issue that it was intended to solve.
arXiv Detail & Related papers (2020-10-12T14:54:32Z) - Learning Bounds for Risk-sensitive Learning [86.50262971918276]
In risk-sensitive learning, one aims to find a hypothesis that minimizes a risk-averse (or risk-seeking) measure of loss.
We study the generalization properties of risk-sensitive learning schemes whose optimand is described via optimized certainty equivalents.
arXiv Detail & Related papers (2020-06-15T05:25:02Z) - Asymptotic normality of robust risk minimizers [2.0432586732993374]
This paper investigates properties of algorithms that can be viewed as robust analogues of the classical empirical risk.
We show that for a wide class of parametric problems, minimizers of the appropriately defined robust proxy of risk converge to the minimizers of the true risk at the same rate.
arXiv Detail & Related papers (2020-04-05T22:03:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.