Spectral density estimation with the Gaussian Integral Transform
- URL: http://arxiv.org/abs/2004.04889v2
- Date: Fri, 14 Aug 2020 17:02:56 GMT
- Title: Spectral density estimation with the Gaussian Integral Transform
- Authors: Alessandro Roggero
- Abstract summary: spectral density operator $hatrho(omega)=delta(omega-hatH)$ plays a central role in linear response theory.
We describe a near optimal quantum algorithm providing an approximation to the spectral density.
- Score: 91.3755431537592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The spectral density operator $\hat{\rho}(\omega)=\delta(\omega-\hat{H})$
plays a central role in linear response theory as its expectation value, the
dynamical response function, can be used to compute scattering cross-sections.
In this work, we describe a near optimal quantum algorithm providing an
approximation to the spectral density with energy resolution $\Delta$ and error
$\epsilon$ using
$\mathcal{O}\left(\sqrt{\log\left(1/\epsilon\right)\left(\log\left(1/\Delta\right)+\log\left(1/\epsilon\right)\right)}/\Delta\right)$
operations. This is achieved without using expensive approximations to the
time-evolution operator but exploiting instead qubitization to implement an
approximate Gaussian Integral Transform (GIT) of the spectral density. We also
describe appropriate error metrics to assess the quality of spectral function
approximations more generally.
Related papers
- WeSpeR: Population spectrum retrieval and spectral density estimation of weighted sample covariance [0.0]
We prove that the spectral distribution $F$ of the weighted sample covariance has a continuous density on $mathbbR*$.
We propose a procedure to compute it, to determine the support of $F$ and define an efficient grid on it.
We use this procedure to design the $textitWeSpeR$ algorithm, which estimates the spectral density and retrieves the true spectral covariance spectrum.
arXiv Detail & Related papers (2024-10-18T12:26:51Z) - Relative-Translation Invariant Wasserstein Distance [82.6068808353647]
We introduce a new family of distances, relative-translation invariant Wasserstein distances ($RW_p$)
We show that $RW_p distances are also real distance metrics defined on the quotient set $mathcalP_p(mathbbRn)/sim$ invariant to distribution translations.
arXiv Detail & Related papers (2024-09-04T03:41:44Z) - Convergence of coordinate ascent variational inference for log-concave measures via optimal transport [0.0]
Mean field inference (VI) is the problem of finding the closest product (factorized) measure.
The well known Ascent Variational Inference (CAVI) aims this approximate measure by variation over one coordinate.
arXiv Detail & Related papers (2024-04-12T19:43:54Z) - An efficient quantum algorithm for generation of ab initio n-th order susceptibilities for non-linear spectroscpies [0.13980986259786224]
We develop and analyze a fault-tolerant quantum algorithm for computing $n$-th order response properties.
We use a semi-classical description in which the electronic degrees of freedom are treated quantum mechanically and the light is treated as a classical field.
arXiv Detail & Related papers (2024-04-01T20:04:49Z) - Optimal and Efficient Algorithms for Decentralized Online Convex Optimization [51.00357162913229]
Decentralized online convex optimization (D-OCO) is designed to minimize a sequence of global loss functions using only local computations and communications.
We develop a novel D-OCO algorithm that can reduce the regret bounds for convex and strongly convex functions to $tildeO(nrho-1/4sqrtT)$ and $tildeO(nrho-1/2log T)$.
Our analysis reveals that the projection-free variant can achieve $O(nT3/4)$ and $O(n
arXiv Detail & Related papers (2024-02-14T13:44:16Z) - Efficient displacement convex optimization with particle gradient
descent [57.88860627977882]
Particle gradient descent is widely used to optimize functions of probability measures.
This paper considers particle gradient descent with a finite number of particles and establishes its theoretical guarantees to optimize functions that are emphdisplacement convex in measures.
arXiv Detail & Related papers (2023-02-09T16:35:59Z) - CEDAS: A Compressed Decentralized Stochastic Gradient Method with Improved Convergence [9.11726703830074]
In this paper, we consider solving the distributed optimization problem under the communication restricted setting.
We show the method over com-pressed exact diffusion termed CEDAS"
In particular, when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when when
arXiv Detail & Related papers (2023-01-14T09:49:15Z) - ReSQueing Parallel and Private Stochastic Convex Optimization [59.53297063174519]
We introduce a new tool for BFG convex optimization (SCO): a Reweighted Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density.
We develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings.
arXiv Detail & Related papers (2023-01-01T18:51:29Z) - Concentration of the Langevin Algorithm's Stationary Distribution [29.244631254978597]
We show that for any nontrivial stepsize $eta > 0$, $pi_eta$ is sub-exponential when the potential is convex.
We also show that these concentration bounds extend to all iterates along the trajectory of the Langevin Algorithm.
arXiv Detail & Related papers (2022-12-24T01:39:02Z) - Quantum state tomography via non-convex Riemannian gradient descent [13.100184125419691]
In this work, we derive a quantum state tomography scheme that improves the dependence on $kappa$ to the scale.
The improvement comes from the application the non-varian gradient (RGD) algorithms.
Our theoretical results of extremely fast convergence and nearly optimal error bounds are corroborated by numerical results.
arXiv Detail & Related papers (2022-10-10T14:19:23Z) - Maximizing the Validity of the Gaussian Approximation for the biphoton
State from Parametric Downconversion [0.0]
We present a choice of $alpha$ which maximizes the validity of the Gaussian approximation.
We also discuss the so-called textitsuper-Gaussian and textitcosine-Gaussian approximations as practical alternatives.
arXiv Detail & Related papers (2022-10-05T15:47:04Z) - Approximate Function Evaluation via Multi-Armed Bandits [51.146684847667125]
We study the problem of estimating the value of a known smooth function $f$ at an unknown point $boldsymbolmu in mathbbRn$, where each component $mu_i$ can be sampled via a noisy oracle.
We design an instance-adaptive algorithm that learns to sample according to the importance of each coordinate, and with probability at least $1-delta$ returns an $epsilon$ accurate estimate of $f(boldsymbolmu)$.
arXiv Detail & Related papers (2022-03-18T18:50:52Z) - Random quantum circuits transform local noise into global white noise [118.18170052022323]
We study the distribution over measurement outcomes of noisy random quantum circuits in the low-fidelity regime.
For local noise that is sufficiently weak and unital, correlations (measured by the linear cross-entropy benchmark) between the output distribution $p_textnoisy$ of a generic noisy circuit instance shrink exponentially.
If the noise is incoherent, the output distribution approaches the uniform distribution $p_textunif$ at precisely the same rate.
arXiv Detail & Related papers (2021-11-29T19:26:28Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.