Bayesian inference of general noise-model parameters from the syndrome statistics of surface codes
- URL: http://arxiv.org/abs/2406.08981v3
- Date: Sun, 19 Oct 2025 23:55:00 GMT
- Title: Bayesian inference of general noise-model parameters from the syndrome statistics of surface codes
- Authors: Takumi Kobori, Synge Todo,
- Abstract summary: Noise model estimation based on syndrome measurement statistics is well-established for Pauli noise.<n>We propose Bayesian inference methods for general noise models, integrating a tensor network simulator of surface code.<n>We present numerical results of applying our proposed methods to various noise models, such as static, time-varying, and nonuniform cases.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The performance of error correction in the surface code can be enhanced by leveraging the knowledge of the noise model for physical qubits. To provide accurate noise information to the decoder in parallel with quantum computation, an adaptive estimation of the noise model based on syndrome measurement statistics is an effective approach. While noise model estimation based on syndrome measurement statistics is well-established for Pauli noise, it remains unexplored for more complex and realistic scenarios such as amplitude damping which cannot be represented as a Pauli channel. In this paper, we propose Bayesian inference methods for general noise models, integrating a tensor network simulator of surface code, which can efficiently simulate various noise models, with Monte Carlo sampling techniques. For stationary noise, we propose a method based on the Markov chain Monte Carlo. For time-varying noise, which is a more realistic scenario, we introduce another method based on the sequential Monte Carlo. We present numerical results of applying our proposed methods to various noise models, such as static, time-varying, and nonuniform cases, and evaluate their performance in detail.
Related papers
- Adaptive Estimation of Drifting Noise in Quantum Error Correction [1.1998722332188005]
We present a framework to capture time-dependent Pauli noise, by exploiting the syndrome statistics of quantum error correction experiments.<n>We prove the noise-filtering behavior of sliding windows, linking window size to spectral cutoff frequencies, and provide an iterative algorithm that captures multiple drift frequencies.<n>Our window-based estimation methods and adaptive decoding offer new insights into noise spectroscopy and decoder optimization under drift.
arXiv Detail & Related papers (2025-11-12T17:03:56Z) - Mitigating the Noise Shift for Denoising Generative Models via Noise Awareness Guidance [54.88271057438763]
Noise Awareness Guidance (NAG) is a correction method that explicitly steers sampling trajectories to remain consistent with the pre-defined noise schedule.<n>NAG consistently mitigates noise shift and substantially improves the generation quality of mainstream diffusion models.
arXiv Detail & Related papers (2025-10-14T13:31:34Z) - Provable Mixed-Noise Learning with Flow-Matching [6.687571760931077]
We study inverse problems with mixed noise, modeled as a combination of additive and multiplicative Gaussian components.<n>Motivated by recent advances in flow-based generative modeling, we propose a novel inference framework based on conditional flow matching embedded within an Expectation-Maximization (EM) algorithm.
arXiv Detail & Related papers (2025-08-25T15:30:12Z) - Diffusion Gaussian Mixture Audio Denoise [23.760755498636943]
We propose a DiffGMM model, a denoising model based on the diffusion and Gaussian mixture models.
Given a noisy audio signal, we first apply a 1D-U-Net to extract features and train linear layers to estimate parameters for the Gaussian mixture model.
The noisy signal is continuously subtracted from the estimated noise to output clean audio signals.
arXiv Detail & Related papers (2024-06-13T14:18:10Z) - Information limits and Thouless-Anderson-Palmer equations for spiked matrix models with structured noise [19.496063739638924]
We consider a saturate problem of Bayesian inference for a structured spiked model.
We show how to predict the statistical limits using an efficient algorithm inspired by the theory of adaptive Thouless-Anderson-Palmer equations.
arXiv Detail & Related papers (2024-05-31T16:38:35Z) - One Noise to Rule Them All: Learning a Unified Model of Spatially-Varying Noise Patterns [33.293193191683145]
We present a single generative model which can learn to generate multiple types of noise as well as blend between them.
We also present an application of our model to improving inverse procedural material design.
arXiv Detail & Related papers (2024-04-25T02:23:11Z) - Blue noise for diffusion models [50.99852321110366]
We introduce a novel and general class of diffusion models taking correlated noise within and across images into account.
Our framework allows introducing correlation across images within a single mini-batch to improve gradient flow.
We perform both qualitative and quantitative evaluations on a variety of datasets using our method.
arXiv Detail & Related papers (2024-02-07T14:59:25Z) - Noisy Pair Corrector for Dense Retrieval [59.312376423104055]
We propose a novel approach called Noisy Pair Corrector (NPC)
NPC consists of a detection module and a correction module.
We conduct experiments on text-retrieval benchmarks Natural Question and TriviaQA, code-search benchmarks StaQC and SO-DS.
arXiv Detail & Related papers (2023-11-07T08:27:14Z) - Volumetric Benchmarking of Quantum Computing Noise Models [3.0098885383612104]
We present a systematic approach to benchmark noise models for quantum computing applications.
It compares the results of hardware experiments to predictions of noise models for a representative set of quantum circuits.
We also construct a noise model and optimize its parameters with a series of training circuits.
arXiv Detail & Related papers (2023-06-14T10:49:01Z) - Realistic Noise Synthesis with Diffusion Models [68.48859665320828]
Deep image denoising models often rely on large amount of training data for the high quality performance.
We propose a novel method that synthesizes realistic noise using diffusion models, namely Realistic Noise Synthesize Diffusor (RNSD)
RNSD can incorporate guided multiscale content, such as more realistic noise with spatial correlations can be generated at multiple frequencies.
arXiv Detail & Related papers (2023-05-23T12:56:01Z) - Improving the Robustness of Summarization Models by Detecting and
Removing Input Noise [50.27105057899601]
We present a large empirical study quantifying the sometimes severe loss in performance from different types of input noise for a range of datasets and model sizes.
We propose a light-weight method for detecting and removing such noise in the input during model inference without requiring any training, auxiliary models, or even prior knowledge of the type of noise.
arXiv Detail & Related papers (2022-12-20T00:33:11Z) - The Optimal Noise in Noise-Contrastive Learning Is Not What You Think [80.07065346699005]
We show that deviating from this assumption can actually lead to better statistical estimators.
In particular, the optimal noise distribution is different from the data's and even from a different family.
arXiv Detail & Related papers (2022-03-02T13:59:20Z) - C2N: Practical Generative Noise Modeling for Real-World Denoising [53.96391787869974]
We introduce a Clean-to-Noisy image generation framework, namely C2N, to imitate complex real-world noise without using paired examples.
We construct the noise generator in C2N accordingly with each component of real-world noise characteristics to express a wide range of noise accurately.
arXiv Detail & Related papers (2022-02-19T05:53:46Z) - Multiview point cloud registration with anisotropic and space-varying
localization noise [1.5499426028105903]
We address the problem of registering multiple point clouds corrupted with high anisotropic localization noise.
Existing methods are based on an implicit assumption of space-invariant isotropic noise.
We show that our noise handling strategy improves significantly the robustness to high levels of anisotropic noise.
arXiv Detail & Related papers (2022-01-03T15:21:24Z) - Noise Estimation for Generative Diffusion Models [91.22679787578438]
In this work, we present a simple and versatile learning scheme that can adjust the noise parameters for any given number of steps.
Our approach comes at a negligible computation cost.
arXiv Detail & Related papers (2021-04-06T15:46:16Z) - Shape Matters: Understanding the Implicit Bias of the Noise Covariance [76.54300276636982]
Noise in gradient descent provides a crucial implicit regularization effect for training over parameterized models.
We show that parameter-dependent noise -- induced by mini-batches or label perturbation -- is far more effective than Gaussian noise.
Our analysis reveals that parameter-dependent noise introduces a bias towards local minima with smaller noise variance, whereas spherical Gaussian noise does not.
arXiv Detail & Related papers (2020-06-15T18:31:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.