Bootstrapping Classical Shadows for Neural Quantum State Tomography
- URL: http://arxiv.org/abs/2405.06864v3
- Date: Wed, 21 Aug 2024 23:42:27 GMT
- Title: Bootstrapping Classical Shadows for Neural Quantum State Tomography
- Authors: Wirawat Kokaew, Bohdan Kulchytskyy, Shunji Matsuura, Pooya Ronagh,
- Abstract summary: We introduce a novel estimator for optimizing the cross-entropy loss function using classical shadows.
We show that this loss function can be used to achieve stable reconstruction of GHZ states using a transformer-based neural network trained on classical shadow measurements.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We investigate the advantages of using autoregressive neural quantum states as ansatze for classical shadow tomography to improve its predictive power. We introduce a novel estimator for optimizing the cross-entropy loss function using classical shadows, and a new importance sampling strategy for estimating the loss gradient during training using stabilizer samples collected from classical shadows. We show that this loss function can be used to achieve stable reconstruction of GHZ states using a transformer-based neural network trained on classical shadow measurements. This loss function also enables the training of neural quantum states representing purifications of mixed states. Our results show that the intrinsic capability of autoregressive models in representing physically well-defined density matrices allows us to overcome the weakness of Pauli-based classical shadow tomography in predicting both high-weight observables and nonlinear observables such as the purity of pure and mixed states.
Related papers
- On the Dynamics Under the Unhinged Loss and Beyond [104.49565602940699]
We introduce the unhinged loss, a concise loss function, that offers more mathematical opportunities to analyze closed-form dynamics.
The unhinged loss allows for considering more practical techniques, such as time-vary learning rates and feature normalization.
arXiv Detail & Related papers (2023-12-13T02:11:07Z) - Neural-Shadow Quantum State Tomography [0.0]
"neural-shadow quantum state tomography" (NSQST) is an alternative neural network-based quantum state tomography protocol.
Infidelity is estimated using the classical shadows of the target state.
NSQST is robust against various types of noise without any error mitigation.
arXiv Detail & Related papers (2023-05-01T20:23:57Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - BiTAT: Neural Network Binarization with Task-dependent Aggregated
Transformation [116.26521375592759]
Quantization aims to transform high-precision weights and activations of a given neural network into low-precision weights/activations for reduced memory usage and computation.
Extreme quantization (1-bit weight/1-bit activations) of compactly-designed backbone architectures results in severe performance degeneration.
This paper proposes a novel Quantization-Aware Training (QAT) method that can effectively alleviate performance degeneration.
arXiv Detail & Related papers (2022-07-04T13:25:49Z) - On Classical and Hybrid Shadows of Quantum States [0.0]
Classical shadows are a computationally efficient approach to storing quantum states on a classical computer.
We discuss the advantages and limitations of using classical shadows to simulate many-body dynamics.
We introduce the notion of a hybrid shadow, constructed from measurements on a part of the system instead of the entirety.
arXiv Detail & Related papers (2022-06-14T06:25:24Z) - On Convergence of Training Loss Without Reaching Stationary Points [62.41370821014218]
We show that Neural Network weight variables do not converge to stationary points where the gradient the loss function vanishes.
We propose a new perspective based on ergodic theory dynamical systems.
arXiv Detail & Related papers (2021-10-12T18:12:23Z) - Classical Shadow Tomography with Locally Scrambled Quantum Dynamics [0.0]
We generalize the classical shadow tomography scheme to a broad class of finite-depth or finite-time local unitary ensembles.
We numerically demonstrate our approach for finite-depth local unitary circuits and finite-time local-Hamiltonian generated evolutions.
arXiv Detail & Related papers (2021-07-10T11:34:51Z) - Adaptive Pauli Shadows for Energy Estimation [0.0]
Recently, derandomised classical shadows have emerged claiming to be even more accurate.
This accuracy comes at a cost of introducing classical computing resources into the energy estimation procedure.
This present note shows, by adding a fraction of this classical computing resource to the locally-biased classical shadows setting, that the modified algorithm, Adaptive Pauli Shadows is state-of-the-art for energy estimation.
arXiv Detail & Related papers (2021-05-25T20:42:35Z) - Gradient Starvation: A Learning Proclivity in Neural Networks [97.02382916372594]
Gradient Starvation arises when cross-entropy loss is minimized by capturing only a subset of features relevant for the task.
This work provides a theoretical explanation for the emergence of such feature imbalance in neural networks.
arXiv Detail & Related papers (2020-11-18T18:52:08Z) - A Bayesian Perspective on Training Speed and Model Selection [51.15664724311443]
We show that a measure of a model's training speed can be used to estimate its marginal likelihood.
We verify our results in model selection tasks for linear models and for the infinite-width limit of deep neural networks.
Our results suggest a promising new direction towards explaining why neural networks trained with gradient descent are biased towards functions that generalize well.
arXiv Detail & Related papers (2020-10-27T17:56:14Z) - Anomaly detection with variational quantum generative adversarial
networks [0.0]
Generative adversarial networks (GANs) are a machine learning framework comprising a generative model for sampling from a target distribution.
We introduce variational quantum-classical Wasserstein GANs to address these issues and embed this model in a classical machine learning framework for anomaly detection.
Our model replaces the generator of Wasserstein GANs with a hybrid quantum-classical neural net and leaves the classical discriminative model unchanged.
arXiv Detail & Related papers (2020-10-20T17:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.