Approximate Dynamics Lead to More Optimal Control: Efficient Exact
Derivatives
- URL: http://arxiv.org/abs/2005.09943v4
- Date: Tue, 18 May 2021 09:47:15 GMT
- Title: Approximate Dynamics Lead to More Optimal Control: Efficient Exact
Derivatives
- Authors: Jesper Hasseriis Mohr Jensen, Frederik Skovbo M{\o}ller, Jens Jakob
S{\o}rensen, Jacob Friis Sherson
- Abstract summary: We show here that the computational feasibility of meeting this accuracy requirement depends on the choice of propagation scheme and problem representation.
This methodology allows numerically efficient optimization of very high-dimensional dynamics.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate derivatives are important for efficiently locally traversing and
converging in quantum optimization landscapes. By deriving analytically exact
control derivatives (gradient and Hessian) for unitary control tasks, we show
here that the computational feasibility of meeting this accuracy requirement
depends on the choice of propagation scheme and problem representation. Even
when exact propagation is sufficiently cheap it is, perhaps surprisingly, much
more efficient to optimize the (appropriately) approximate propagators:
approximations in the dynamics are traded off for significant complexity
reductions in the exact derivative calculations. Importantly, past the initial
analytical considerations, only standard numerical techniques are explicitly
required with straightforward application to realistic systems. These results
are numerically verified for two concrete problems of increasing Hilbert space
dimensionality. The best schemes obtain unit fidelity to machine precision
whereas the results for other schemes are separated consistently by orders of
magnitude in computation time and in worst case 10 orders of magnitude in
achievable fidelity. Since these gaps continually increase with system size and
complexity, this methodology allows numerically efficient optimization of very
high-dimensional dynamics, e.g. in many-body contexts, operating in the
high-fidelity regime which will be published separately.
Related papers
- WENDy for Nonlinear-in-Parameters ODEs [1.9573380763700712]
The Weak-form Estimation of Non-linear Dynamics (WEN) is extended to accommodate systems of ordinary differential equations that are nonlinear-ins.
We present results on a suite of benchmark systems to demonstrate the practical benefits of our approach.
arXiv Detail & Related papers (2025-02-13T01:40:21Z) - PINS: Proximal Iterations with Sparse Newton and Sinkhorn for Optimal Transport [12.551419304993209]
We introduce Proximal Iterations with Sparse Newton and Sinkhorn methods to efficiently compute highly accurate solutions for large-scale OT problems.
A reduced computational complexity through overall sparsity and global convergence are guaranteed by rigorous theoretical analysis.
Our approach offers three key advantages: it achieves accuracy comparable to exact solutions, progressively accelerates each iteration for greater efficiency, and enhances robustness by reducing sensitivity to regularization parameters.
arXiv Detail & Related papers (2025-02-06T03:21:48Z) - Understanding Optimization in Deep Learning with Central Flows [53.66160508990508]
We show that an RMS's implicit behavior can be explicitly captured by a "central flow:" a differential equation.
We show that these flows can empirically predict long-term optimization trajectories of generic neural networks.
arXiv Detail & Related papers (2024-10-31T17:58:13Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Generative modeling of time-dependent densities via optimal transport
and projection pursuit [3.069335774032178]
We propose a cheap alternative to popular deep learning algorithms for temporal modeling.
Our method is highly competitive compared with state-of-the-art solvers.
arXiv Detail & Related papers (2023-04-19T13:50:13Z) - Neural Solvers for Fast and Accurate Numerical Optimal Control [12.80824586913772]
This paper provides techniques to improve the quality of optimized control policies given a fixed computational budget.
We achieve the above via a hypersolvers approach, which hybridizes a differential equation solver and a neural network.
arXiv Detail & Related papers (2022-03-13T10:46:50Z) - Fast Distributionally Robust Learning with Variance Reduced Min-Max
Optimization [85.84019017587477]
Distributionally robust supervised learning is emerging as a key paradigm for building reliable machine learning systems for real-world applications.
Existing algorithms for solving Wasserstein DRSL involve solving complex subproblems or fail to make use of gradients.
We revisit Wasserstein DRSL through the lens of min-max optimization and derive scalable and efficiently implementable extra-gradient algorithms.
arXiv Detail & Related papers (2021-04-27T16:56:09Z) - Self Normalizing Flows [65.73510214694987]
We propose a flexible framework for training normalizing flows by replacing expensive terms in the gradient by learned approximate inverses at each layer.
This reduces the computational complexity of each layer's exact update from $mathcalO(D3)$ to $mathcalO(D2)$.
We show experimentally that such models are remarkably stable and optimize to similar data likelihood values as their exact gradient counterparts.
arXiv Detail & Related papers (2020-11-14T09:51:51Z) - Efficient Learning of Generative Models via Finite-Difference Score
Matching [111.55998083406134]
We present a generic strategy to efficiently approximate any-order directional derivative with finite difference.
Our approximation only involves function evaluations, which can be executed in parallel, and no gradient computations.
arXiv Detail & Related papers (2020-07-07T10:05:01Z) - Real-Time Regression with Dividing Local Gaussian Processes [62.01822866877782]
Local Gaussian processes are a novel, computationally efficient modeling approach based on Gaussian process regression.
Due to an iterative, data-driven division of the input space, they achieve a sublinear computational complexity in the total number of training points in practice.
A numerical evaluation on real-world data sets shows their advantages over other state-of-the-art methods in terms of accuracy as well as prediction and update speed.
arXiv Detail & Related papers (2020-06-16T18:43:31Z) - Beyond the Mean-Field: Structured Deep Gaussian Processes Improve the
Predictive Uncertainties [12.068153197381575]
We propose a novel variational family that allows for retaining covariances between latent processes while achieving fast convergence.
We provide an efficient implementation of our new approach and apply it to several benchmark datasets.
It yields excellent results and strikes a better balance between accuracy and calibrated uncertainty estimates than its state-of-the-art alternatives.
arXiv Detail & Related papers (2020-05-22T11:10:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.