Operator convexity along lines, self-concordance, and sandwiched Rényi entropies
- URL: http://arxiv.org/abs/2502.05627v1
- Date: Sat, 08 Feb 2025 16:17:57 GMT
- Title: Operator convexity along lines, self-concordance, and sandwiched Rényi entropies
- Authors: Kerry He, James Saunderson, Hamza Fawzi,
- Abstract summary: We show that if a convex function is operator convex along any one-dimensional restriction, the natural logarithmic barrier of its epigraph is self-concordant.
An implementation of the convex cones considered in this paper is now available in our open source interior-point solver QICS.
- Score: 6.281229317487581
- License:
- Abstract: Barrier methods play a central role in the theory and practice of convex optimization. One of the most general and successful analyses of barrier methods for convex optimization, due to Nesterov and Nemirovskii, relies on the notion of self-concordance. While an extremely powerful concept, proving self-concordance of barrier functions can be very difficult. In this paper we give a simple way to verify that the natural logarithmic barrier of a convex nonlinear constraint is self-concordant via the theory of operator convex functions. Namely, we show that if a convex function is operator convex along any one-dimensional restriction, then the natural logarithmic barrier of its epigraph is self-concordant. We apply this technique to construct self-concordant barriers for the epigraphs of functions arising in quantum information theory. Notably, we apply this to the sandwiched R\'enyi entropy function, for which no self-concordant barrier was known before. Additionally, we utilize our sufficient condition to provide simplified proofs for previously established self-concordance results for the noncommutative perspective of operator convex functions. An implementation of the convex cones considered in this paper is now available in our open source interior-point solver QICS.
Related papers
- Tightening convex relaxations of trained neural networks: a unified approach for convex and S-shaped activations [0.0]
We develop a formula that yields a tight convexification for the composition of an activation with an affine function.
Our approach can be used to efficiently compute separating hyperplanes or determine that none exists in various settings.
arXiv Detail & Related papers (2024-10-30T18:09:53Z) - Disciplined Geodesically Convex Programming [0.9899763598214121]
Testing convexity structure in nonlinear programs relies on verifying the convexity objectives and constraints.
citetgrantdisciplined introduced a framework, Disciplined Convex Programming (DCP), which can verify basic convexatoms.
Our paper is accompanied by a Julia package, which provides functionality for testing DGCP-compliant expressions.
arXiv Detail & Related papers (2024-07-07T05:13:51Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Constrained Optimization of Rank-One Functions with Indicator Variables [0.0]
optimization problems involving a rank-one convex function over constraints modeling restrictions on the support of the decision variables emerge in various machine learning applications.
We propose a constructive approach that exploits a hidden conic structure induced by perspective functions.
This enables us to systematically give perspective formulations for the convex hull descriptions of sets with nonlinear separable or non-separable objective functions, sign constraints on continuous variables, and constraints on indicator variables.
arXiv Detail & Related papers (2023-03-31T15:51:56Z) - Third quantization of open quantum systems: new dissipative symmetries
and connections to phase-space and Keldysh field theory formulations [77.34726150561087]
We reformulate the technique of third quantization in a way that explicitly connects all three methods.
We first show that our formulation reveals a fundamental dissipative symmetry present in all quadratic bosonic or fermionic Lindbladians.
For bosons, we then show that the Wigner function and the characteristic function can be thought of as ''wavefunctions'' of the density matrix.
arXiv Detail & Related papers (2023-02-27T18:56:40Z) - Confidence Sets under Generalized Self-Concordance [2.0305676256390934]
This paper revisits a fundamental problem in statistical from a non-asymptotic theoretical viewpoint.
We establish an exponential-bound for the estimator characterizing its behavior in a non-asymptotic fashion.
An important trace of its dependency is captured by the effective dimension.
arXiv Detail & Related papers (2022-12-31T17:45:11Z) - Optimal self-concordant barriers for quantum relative entropies [2.1574781022415364]
We prove self-concordance of natural barrier functions for the epigraphs of various quantum relative entropies and divergences.
These barriers allow convex optimization problems involving quantum relative entropies to be directly solved.
arXiv Detail & Related papers (2022-05-09T22:06:18Z) - Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency [111.83670279016599]
We study reinforcement learning for partially observed decision processes (POMDPs) with infinite observation and state spaces.
We make the first attempt at partial observability and function approximation for a class of POMDPs with a linear structure.
arXiv Detail & Related papers (2022-04-20T21:15:38Z) - Learning PSD-valued functions using kernel sums-of-squares [94.96262888797257]
We introduce a kernel sum-of-squares model for functions that take values in the PSD cone.
We show that it constitutes a universal approximator of PSD functions, and derive eigenvalue bounds in the case of subsampled equality constraints.
We then apply our results to modeling convex functions, by enforcing a kernel sum-of-squares representation of their Hessian.
arXiv Detail & Related papers (2021-11-22T16:07:50Z) - On dissipative symplectic integration with applications to
gradient-based optimization [77.34726150561087]
We propose a geometric framework in which discretizations can be realized systematically.
We show that a generalization of symplectic to nonconservative and in particular dissipative Hamiltonian systems is able to preserve rates of convergence up to a controlled error.
arXiv Detail & Related papers (2020-04-15T00:36:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.