$α$-$z$-Rényi divergences in von Neumann algebras: data-processing inequality, reversibility, and monotonicity properties in $α,z$
- URL: http://arxiv.org/abs/2404.07617v2
- Date: Mon, 30 Sep 2024 06:33:47 GMT
- Title: $α$-$z$-Rényi divergences in von Neumann algebras: data-processing inequality, reversibility, and monotonicity properties in $α,z$
- Authors: Fumio Hiai, Anna Jenčová,
- Abstract summary: We prove the variational expressions and the data processing inequality (DPI) for the $alpha$-$z$-R'enyi divergences.
We show the monotonicity properties of $D_alpha,z(psi|varphi)$ in the parameters $alpha,z$ and their limits to the normalized relative entropy.
- Score: 2.3020018305241337
- License:
- Abstract: We study the $\alpha$-$z$-R\'enyi divergences $D_{\alpha,z}(\psi\|\varphi)$ where $\alpha,z>0$ ($\alpha\ne1$) for normal positive functionals $\psi,\varphi$ on general von Neumann algebras, introduced in [S.~Kato and Y.~Ueda, arXiv:2307.01790] and [S.~Kato, arXiv:2311.01748]. We prove the variational expressions and the data processing inequality (DPI) for the $\alpha$-$z$-R\'enyi divergences. We establish the sufficiency theorem for $D_{\alpha,z}(\psi\|\varphi)$, saying that for $(\alpha,z)$ inside the DPI bounds, the equality $D_{\alpha,z}(\psi\circ\gamma\|\varphi\circ\gamma)=D_{\alpha,z}(\psi\|\varphi)<\infty$ in the DPI under a quantum channel (or a normal $2$-positive unital map) $\gamma$ implies the reversibility of $\gamma$ with respect to $\psi,\varphi$. Moreover, we show the monotonicity properties of $D_{\alpha,z}(\psi\|\varphi)$ in the parameters $\alpha,z$ and their limits to the normalized relative entropy as $\alpha\nearrow1$ and $\alpha\searrow1$.
Related papers
- LevAttention: Time, Space, and Streaming Efficient Algorithm for Heavy Attentions [54.54897832889028]
We show that for any $K$, there is a universal set" $U subset [n]$ of size independent of $n$, such that for any $Q$ and any row $i$, the large attention scores $A_i,j$ in row $i$ of $A$ all have $jin U$.
We empirically show the benefits of our scheme for vision transformers, showing how to train new models that use our universal set while training as well.
arXiv Detail & Related papers (2024-10-07T19:47:13Z) - $\ell_p$-Regression in the Arbitrary Partition Model of Communication [59.89387020011663]
We consider the randomized communication complexity of the distributed $ell_p$-regression problem in the coordinator model.
For $p = 2$, i.e., least squares regression, we give the first optimal bound of $tildeTheta(sd2 + sd/epsilon)$ bits.
For $p in (1,2)$,we obtain an $tildeO(sd2/epsilon + sd/mathrmpoly(epsilon)$ upper bound.
arXiv Detail & Related papers (2023-07-11T08:51:53Z) - Average R\'{e}nyi Entropy of a Subsystem in Random Pure State [0.0]
We plot the $ln m$-dependence of the quantum information derived from $widetildeS_alpha (m,n)$.
It is remarkable to note that the nearly vanishing region of the information becomes shorten with increasing $alpha$, and eventually disappears in the limit of $alpha rightarrow infty$.
arXiv Detail & Related papers (2023-01-22T08:08:51Z) - Bounds on Renyi entropy growth in many-body quantum systems [0.0]
We prove rigorous bounds on the growth of $alpha$-Renyi entropies $S_alpha(t)$.
For completely non-local Hamiltonians, we show that the instantaneous growth rates $|S'_alpha(t)|$ can be exponentially larger than $|S'_alpha(t)|$.
arXiv Detail & Related papers (2022-12-14T19:00:01Z) - Quantum Entanglement with Generalized Uncertainty Principle [0.0]
We explore how the quantum entanglement is modified in the generalized uncertainty principle (GUP)-corrected quantum mechanics.
It is shown that $cal E_gamma (rho_A)$ increases with increasing $alpha$ when $gamma = 2, 3, cdots$.
The remarkable fact is that $cal E_EoF (rho_A)$ does not have first-order of $alpha$.
arXiv Detail & Related papers (2022-03-13T03:25:46Z) - Low-Rank Approximation with $1/\epsilon^{1/3}$ Matrix-Vector Products [58.05771390012827]
We study iterative methods based on Krylov subspaces for low-rank approximation under any Schatten-$p$ norm.
Our main result is an algorithm that uses only $tildeO(k/sqrtepsilon)$ matrix-vector products.
arXiv Detail & Related papers (2022-02-10T16:10:41Z) - Active Sampling for Linear Regression Beyond the $\ell_2$ Norm [70.49273459706546]
We study active sampling algorithms for linear regression, which aim to query only a small number of entries of a target vector.
We show that this dependence on $d$ is optimal, up to logarithmic factors.
We also provide the first total sensitivity upper bound $O(dmax1,p/2log2 n)$ for loss functions with at most degree $p$ growth.
arXiv Detail & Related papers (2021-11-09T00:20:01Z) - Learning low-degree functions from a logarithmic number of random
queries [77.34726150561087]
We prove that for any integer $ninmathbbN$, $din1,ldots,n$ and any $varepsilon,deltain(0,1)$, a bounded function $f:-1,1nto[-1,1]$ of degree at most $d$ can be learned.
arXiv Detail & Related papers (2021-09-21T13:19:04Z) - Computation of sandwiched relative alpha-entropy of two n-mode gaussian
states [0.0]
formula for sandwiched relative $alpha$-entropy $widetildeD_alpha(rhovertvertsigma)=frac1alpha-1, ln,rm Tr, left(sigmafrac1-alpha2alpha,rho,sigmafrac1-alpha2alpharight)alpha for $0 alpha 1$, of two $n$ mode rhogaussian states $mathbbCn
arXiv Detail & Related papers (2021-09-09T07:00:30Z) - Spectral properties of sample covariance matrices arising from random
matrices with independent non identically distributed columns [50.053491972003656]
It was previously shown that the functionals $texttr(AR(z))$, for $R(z) = (frac1nXXT- zI_p)-1$ and $Ain mathcal M_p$ deterministic, have a standard deviation of order $O(|A|_* / sqrt n)$.
Here, we show that $|mathbb E[R(z)] - tilde R(z)|_F
arXiv Detail & Related papers (2021-09-06T14:21:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.