Accelerating Continuous Variable Coherent Ising Machines via Momentum
- URL: http://arxiv.org/abs/2401.12135v1
- Date: Mon, 22 Jan 2024 17:18:53 GMT
- Title: Accelerating Continuous Variable Coherent Ising Machines via Momentum
- Authors: Robin Brown, Davide Venturelli, Marco Pavone, and David E. Bernal
Neira
- Abstract summary: We propose to modify CV-CIM dynamics using more tunable optimization techniques such as momentum and Adam.
We show that momentum and Adam-CIM's and sample Adam-CV-CIM's performance is more stable as an tunable framework.
- Score: 16.545815849819043
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Coherent Ising Machine (CIM) is a non-conventional architecture that
takes inspiration from physical annealing processes to solve Ising problems
heuristically. Its dynamics are naturally continuous and described by a set of
ordinary differential equations that have been proven to be useful for the
optimization of continuous variables non-convex quadratic optimization
problems. The dynamics of such Continuous Variable CIMs (CV-CIM) encourage
optimization via optical pulses whose amplitudes are determined by the negative
gradient of the objective; however, standard gradient descent is known to be
trapped by local minima and hampered by poor problem conditioning. In this
work, we propose to modify the CV-CIM dynamics using more sophisticated pulse
injections based on tried-and-true optimization techniques such as momentum and
Adam. Through numerical experiments, we show that the momentum and Adam updates
can significantly speed up the CV-CIM's convergence and improve sample
diversity over the original CV-CIM dynamics. We also find that the
Adam-CV-CIM's performance is more stable as a function of feedback strength,
especially on poorly conditioned instances, resulting in an algorithm that is
more robust, reliable, and easily tunable. More broadly, we identify the CIM
dynamical framework as a fertile opportunity for exploring the intersection of
classical optimization and modern analog computing.
Related papers
- Losing momentum in continuous-time stochastic optimisation [42.617042045455506]
momentum-based optimisation algorithms have become particularly widespread.
In this work, we analyse a continuous-time model for gradient descent with momentum.
We also train a convolutional neural network in an image classification problem.
arXiv Detail & Related papers (2022-09-08T10:46:05Z) - On Fast Simulation of Dynamical System with Neural Vector Enhanced
Numerical Solver [59.13397937903832]
We introduce a deep learning-based corrector called Neural Vector (NeurVec)
NeurVec can compensate for integration errors and enable larger time step sizes in simulations.
Our experiments on a variety of complex dynamical system benchmarks demonstrate that NeurVec exhibits remarkable generalization capability.
arXiv Detail & Related papers (2022-08-07T09:02:18Z) - Learning Sparse Nonlinear Dynamics via Mixed-Integer Optimization [3.7565501074323224]
We propose an exact formulation of the SINDyDy problem using mixed-integer optimization (MIO) to solve the sparsity constrained regression problem to provable optimality in seconds.
We illustrate the dramatic improvement of our approach in accurate model discovery while being more sample efficient, robust to noise, and flexible in accommodating physical constraints.
arXiv Detail & Related papers (2022-06-01T01:43:45Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - Robust Value Iteration for Continuous Control Tasks [99.00362538261972]
When transferring a control policy from simulation to a physical system, the policy needs to be robust to variations in the dynamics to perform well.
We present Robust Fitted Value Iteration, which uses dynamic programming to compute the optimal value function on the compact state domain.
We show that robust value is more robust compared to deep reinforcement learning algorithm and the non-robust version of the algorithm.
arXiv Detail & Related papers (2021-05-25T19:48:35Z) - Value Iteration in Continuous Actions, States and Time [99.00362538261972]
We propose a continuous fitted value iteration (cFVI) algorithm for continuous states and actions.
The optimal policy can be derived for non-linear control-affine dynamics.
Videos of the physical system are available at urlhttps://sites.google.com/view/value-iteration.
arXiv Detail & Related papers (2021-05-10T21:40:56Z) - EBMs Trained with Maximum Likelihood are Generator Models Trained with a
Self-adverserial Loss [6.445605125467574]
We replace Langevin dynamics with deterministic solutions of the associated gradient descent ODE.
We show that reintroducing the noise in the dynamics does not lead to a qualitative change in the behavior.
We thus show that EBM training is effectively a self-adversarial procedure rather than maximum likelihood estimation.
arXiv Detail & Related papers (2021-02-23T15:34:12Z) - Training Generative Adversarial Networks by Solving Ordinary
Differential Equations [54.23691425062034]
We study the continuous-time dynamics induced by GAN training.
From this perspective, we hypothesise that instabilities in training GANs arise from the integration error.
We experimentally verify that well-known ODE solvers (such as Runge-Kutta) can stabilise training.
arXiv Detail & Related papers (2020-10-28T15:23:49Z) - Quasi-symplectic Langevin Variational Autoencoder [7.443843354775884]
Variational autoencoder (VAE) is a very popular and well-investigated generative model in neural learning research.
It is required to deal with the difficulty of building low variance evidence lower bounds (ELBO)
arXiv Detail & Related papers (2020-09-02T12:13:27Z) - Multiplicative noise and heavy tails in stochastic optimization [62.993432503309485]
empirical optimization is central to modern machine learning, but its role in its success is still unclear.
We show that it commonly arises in parameters of discrete multiplicative noise due to variance.
A detailed analysis is conducted in which we describe on key factors, including recent step size, and data, all exhibit similar results on state-of-the-art neural network models.
arXiv Detail & Related papers (2020-06-11T09:58:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.