Projective squeezing for translation symmetric bosonic codes
- URL: http://arxiv.org/abs/2403.14218v1
- Date: Thu, 21 Mar 2024 08:19:47 GMT
- Title: Projective squeezing for translation symmetric bosonic codes
- Authors: Suguru Endo, Keitaro Anai, Yuichiro Matsuzaki, Yuuki Tokunaga, Yasunari Suzuki,
- Abstract summary: We introduce the textitprojective squeezing (PS) method for computing outcomes for a higher squeezing level.
We numerically verify our analytical arguments and show that our protocol can mitigate the effect of photon loss.
- Score: 0.16777183511743468
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The design of translation symmetric bosonic codes, e.g., Gottesmann-Kitaev-Preskill and squeezed cat codes, is robust against photon loss, but the computation accuracy is limited by the available squeezing level. Here, we introduce the \textit{projective squeezing} (PS) method for computing outcomes for a higher squeezing level by revealing that a linear combination of displacement operators with periodic displacement values constitutes the smeared projector onto the better code space; we also show the analytical relationship between the increased squeezing level and the projection probability. We introduce concrete implementation methods for PS based on linear-combination-of-unitaries and virtual quantum error detection. We also numerically verify our analytical arguments and show that our protocol can mitigate the effect of photon loss.
Related papers
- The END: An Equivariant Neural Decoder for Quantum Error Correction [73.4384623973809]
We introduce a data efficient neural decoder that exploits the symmetries of the problem.
We propose a novel equivariant architecture that achieves state of the art accuracy compared to previous neural decoders.
arXiv Detail & Related papers (2023-04-14T19:46:39Z) - Regularized Vector Quantization for Tokenized Image Synthesis [126.96880843754066]
Quantizing images into discrete representations has been a fundamental problem in unified generative modeling.
deterministic quantization suffers from severe codebook collapse and misalignment with inference stage while quantization suffers from low codebook utilization and reconstruction objective.
This paper presents a regularized vector quantization framework that allows to mitigate perturbed above issues effectively by applying regularization from two perspectives.
arXiv Detail & Related papers (2023-03-11T15:20:54Z) - Experimental realization of deterministic and selective photon addition
in a bosonic mode assisted by an ancillary qubit [50.591267188664666]
Bosonic quantum error correcting codes are primarily designed to protect against single-photon loss.
Error correction requires a recovery operation that maps the error states -- which have opposite parity -- back onto the code states.
Here, we realize a collection of photon-number-selective, simultaneous photon addition operations on a bosonic mode.
arXiv Detail & Related papers (2022-12-22T23:32:21Z) - Importance sampling for stochastic quantum simulations [68.8204255655161]
We introduce the qDrift protocol, which builds random product formulas by sampling from the Hamiltonian according to the coefficients.
We show that the simulation cost can be reduced while achieving the same accuracy, by considering the individual simulation cost during the sampling stage.
Results are confirmed by numerical simulations performed on a lattice nuclear effective field theory.
arXiv Detail & Related papers (2022-12-12T15:06:32Z) - Parity-encoding-based quantum computing with Bayesian error tracking [0.0]
Measurement-based quantum computing (MBQC) in linear optical systems is promising for near-future quantum computing architecture.
We propose a linear optical topological MBQC protocol employing multiphoton qubits based on the parity encoding.
We show that our protocol is advantageous over several other existing approaches in terms of fault-tolerance, resource overhead, or feasibility of basic elements.
arXiv Detail & Related papers (2022-07-14T10:32:05Z) - Performance of teleportation-based error correction circuits for bosonic
codes with noisy measurements [58.720142291102135]
We analyze the error-correction capabilities of rotation-symmetric codes using a teleportation-based error-correction circuit.
We find that with the currently achievable measurement efficiencies in microwave optics, bosonic rotation codes undergo a substantial decrease in their break-even potential.
arXiv Detail & Related papers (2021-08-02T16:12:13Z) - Gaussian boson sampling with partial distinguishability [0.0]
We investigate GBS with partial distinguishability using an approach based on virtual modes and indistinguishability efficiency.
We show how the boundary of quantum supremacy in GBS can be pushed further by partial distinguishability.
arXiv Detail & Related papers (2021-05-20T08:17:51Z) - Learned transform compression with optimized entropy encoding [72.20409648915398]
We consider the problem of learned transform compression where we learn both, the transform and the probability distribution over the discrete codes.
We employ a soft relaxation of the quantization operation to allow for back-propagation of gradients and employ vector (rather than scalar) quantization of the latent codes.
arXiv Detail & Related papers (2021-04-07T17:58:01Z) - Fault-tolerant quantum computation with static linear optics [0.0]
In this work we propose a topologically error-corrected architecture that does away with these elements at no cost.
Our computer consists of three modules: a 2D array of probabilistic sources of GKP states; a depth-four circuit of static beamsplitters, phase shifters, and single-time-step delay lines.
The symmetry of our proposed circuit allows us to combine the effects of finite squeezing and uniform photon loss within the noise model, resulting in more comprehensive threshold estimates.
arXiv Detail & Related papers (2021-04-07T16:43:34Z) - One-Bit Compressed Sensing via One-Shot Hard Thresholding [7.594050968868919]
A problem of 1-bit compressed sensing is to estimate a sparse signal from a few binary measurements.
We present a novel and concise analysis that moves away from the widely used non-constrained notion of width.
arXiv Detail & Related papers (2020-07-07T17:28:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.