Hardware Co-Designed Optimal Control for Programmable Atomic Quantum Processors via Reinforcement Learning
- URL: http://arxiv.org/abs/2504.11737v1
- Date: Wed, 16 Apr 2025 03:30:40 GMT
- Title: Hardware Co-Designed Optimal Control for Programmable Atomic Quantum Processors via Reinforcement Learning
- Authors: Qian Ding, Dirk Englund,
- Abstract summary: We introduce a hardware co-designed quantum control framework to address inherent imperfections in classical control hardware.<n>We demonstrate that the proposed framework enables robust, high-fidelity parallel single-qubit gate operations.<n>We find that while PPO performance degrades as system complexity increases, the end-to-end differentiable RL consistently achieves gate fidelities above 99.9$%$.
- Score: 0.18416014644193068
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Developing scalable, fault-tolerant atomic quantum processors requires precise control over large arrays of optical beams. This remains a major challenge due to inherent imperfections in classical control hardware, such as inter-channel crosstalk and beam leakage. In this work, we introduce a hardware co-designed intelligent quantum control framework to address these limitations. We construct a mathematical model of the photonic control hardware, integrate it into the quantum optimal control (QOC) framework, and apply reinforcement learning (RL) techniques to discover optimal control strategies. We demonstrate that the proposed framework enables robust, high-fidelity parallel single-qubit gate operations under realistic control conditions, where each atom is individually addressed by an optical beam. Specifically, we implement and benchmark three optimization strategies: a classical hybrid Self-Adaptive Differential Evolution-Adam (SADE-Adam) optimizer, a conventional RL approach based on Proximal Policy Optimization (PPO), and a novel end-to-end differentiable RL method. Using SADE-Adam as a baseline, we find that while PPO performance degrades as system complexity increases, the end-to-end differentiable RL consistently achieves gate fidelities above 99.9$\%$, exhibits faster convergence, and maintains robustness under varied channel crosstalk strength and randomized dynamic control imperfections.
Related papers
- Reinforcement Learning for Quantum Control under Physical Constraints [2.874893537471256]
We devise a physics-informed Reinforcement Learning algorithm that restricts the space of possible solutions.
We evaluate our method on three broadly relevant quantum systems.
arXiv Detail & Related papers (2025-01-24T10:11:32Z) - Enhancing Spectrum Efficiency in 6G Satellite Networks: A GAIL-Powered Policy Learning via Asynchronous Federated Inverse Reinforcement Learning [67.95280175998792]
A novel adversarial imitation learning (GAIL)-powered policy learning approach is proposed for optimizing beamforming, spectrum allocation, and remote user equipment (RUE) association ins.
We employ inverse RL (IRL) to automatically learn reward functions without manual tuning.
We show that the proposed MA-AL method outperforms traditional RL approaches, achieving a $14.6%$ improvement in convergence and reward value.
arXiv Detail & Related papers (2024-09-27T13:05:02Z) - Controlling nonergodicity in quantum many-body systems by reinforcement learning [0.0]
We develop a model-free and deep-reinforcement learning framework for quantum nonergodicity control.<n>We use the paradigmatic one-dimensional tilted Fermi-Hubbard system to demonstrate that the DRL agent can efficiently learn the quantum many-body system.<n>The continuous control protocols and observations are experimentally feasible.
arXiv Detail & Related papers (2024-08-21T20:55:44Z) - Machine-learning-inspired quantum optimal control of nonadiabatic
geometric quantum computation via reverse engineering [3.3216171033358077]
We propose a promising average-fidelity-based machine-learning-inspired method to optimize the control parameters.
We implement a single-qubit gate by cat-state nonadiabatic geometric quantum computation via reverse engineering.
We demonstrate that the neural network possesses the ability to expand the model space.
arXiv Detail & Related papers (2023-09-28T14:36:26Z) - Optimizing quantum gates towards the scale of logical qubits [78.55133994211627]
A foundational assumption of quantum gates theory is that quantum gates can be scaled to large processors without exceeding the error-threshold for fault tolerance.
Here we report on a strategy that can overcome such problems.
We demonstrate it by choreographing the frequency trajectories of 68 frequency-tunablebits to execute single qubit while superconducting errors.
arXiv Detail & Related papers (2023-08-04T13:39:46Z) - Quantum Gate Optimization for Rydberg Architectures in the Weak-Coupling
Limit [55.05109484230879]
We demonstrate machine learning assisted design of a two-qubit gate in a Rydberg tweezer system.
We generate optimal pulse sequences that implement a CNOT gate with high fidelity.
We show that local control of single qubit operations is sufficient for performing quantum computation on a large array of atoms.
arXiv Detail & Related papers (2023-06-14T18:24:51Z) - Direct pulse-level compilation of arbitrary quantum logic gates on superconducting qutrits [36.30869856057226]
We demonstrate any arbitrary qubit and qutrit gate can be realized with high-fidelity, which can significantly reduce the length of a gate sequence.
We show that optimal control gates are robust to drift for at least three hours and that the same calibration parameters can be used for all implemented gates.
arXiv Detail & Related papers (2023-03-07T22:15:43Z) - Optimal solutions to quantum annealing using two independent control
functions [0.0]
We show that an optimal solution consists of both controls tuned at their upper bound for the whole evolution time.
We propose the use of a quantum optimal control technique adapted to limit the amplitude of the controls.
We show that the scheme with two-control functions yields a higher fidelity than the other schemes for the same evolution time.
arXiv Detail & Related papers (2021-10-26T16:54:17Z) - Noise-Robust End-to-End Quantum Control using Deep Autoregressive Policy
Networks [2.5946789143276447]
Variational quantum eigensolvers have recently received increased attention, as they enable the use of quantum computing devices.
We present a hybrid policy gradient algorithm capable of simultaneously optimizing continuous and discrete degrees of freedom in an uncertainty-resilient way.
Our work exhibits the beneficial synergy between reinforcement learning and quantum control.
arXiv Detail & Related papers (2020-12-12T02:13:28Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - Optimization-driven Deep Reinforcement Learning for Robust Beamforming
in IRS-assisted Wireless Communications [54.610318402371185]
Intelligent reflecting surface (IRS) is a promising technology to assist downlink information transmissions from a multi-antenna access point (AP) to a receiver.
We minimize the AP's transmit power by a joint optimization of the AP's active beamforming and the IRS's passive beamforming.
We propose a deep reinforcement learning (DRL) approach that can adapt the beamforming strategies from past experiences.
arXiv Detail & Related papers (2020-05-25T01:42:55Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.