Accelerated characterization of two-level systems in superconducting qubits via machine learning
- URL: http://arxiv.org/abs/2509.17723v1
- Date: Mon, 22 Sep 2025 12:55:52 GMT
- Title: Accelerated characterization of two-level systems in superconducting qubits via machine learning
- Authors: Avinash Pathapati, Olli Mansikkamäki, Alexander Tyner, Alexander V. Balatsky,
- Abstract summary: We introduce a data-driven approach for extracting two-level system (TLS) parameters-frequency $omega_TLS$, coupling strength $g$, dissipation time $T_TLS, 1$, and the pure dephasing time $Tphi_TLS, 2$.<n>A custom convolutional neural network model(CNN) can simultaneously predict $omega_TLS$, $g$, $T_TLS, 1$ and $Tphi_TLS, 2$ from the spectroscopy data presented in the form of
- Score: 74.02542457579791
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a data-driven approach for extracting two-level system (TLS) parameters-frequency $\omega_{TLS}$, coupling strength $g$, dissipation time $T_{TLS, 1}$, and the pure dephasing time $T^{\phi}_{TLS, 2}$, labelled as a 4-component vector $\vec{q}$, directly from simulated spectroscopy data generated for a single TLS by a form of two-tone spectroscopy. Specifically, we demonstrate that a custom convolutional neural network model(CNN) can simultaneously predict $\omega_{TLS}$, $g$, $T_{TLS, 1}$ and $T^{\phi}_{TLS, 2}$ from the spectroscopy data presented in the form of images. Our results show that the model achieves superior performance to perturbation theory methods in successfully extracting the TLS parameters. Although the model, initially trained on noise-free data, exhibits a decline in accuracy when evaluated on noisy images, retraining it on a noisy dataset leads to a substantial performance improvement, achieving results comparable to those obtained under noise-free conditions. Furthermore, the model exhibits higher predictive accuracy for parameters $\omega_{TLS}$ and $g$ in comparison to $T_{TLS, 1}$ and $T^{\phi}_{TLS, 2}$.
Related papers
- Neural Networks Learn Generic Multi-Index Models Near Information-Theoretic Limit [66.20349460098275]
We study the gradient descent learning of a general Gaussian Multi-index model $f(boldsymbolx)=g(boldsymbolUboldsymbolx)$ with hidden subspace $boldsymbolUin mathbbRrtimes d$.<n>We prove that under generic non-degenerate assumptions on the link function, a standard two-layer neural network trained via layer-wise gradient descent can agnostically learn the target with $o_d(1)$ test error.
arXiv Detail & Related papers (2025-11-19T04:46:47Z) - Test time training enhances in-context learning of nonlinear functions [51.56484100374058]
Test-time training (TTT) enhances model performance by explicitly updating designated parameters prior to each prediction.<n>We investigate the combination of TTT with in-context learning (ICL), where the model is given a few examples from the target distribution at inference time.
arXiv Detail & Related papers (2025-09-30T03:56:44Z) - Exploring Microstructural Dynamics in Cryptocurrency Limit Order Books: Better Inputs Matter More Than Stacking Another Hidden Layer [9.2463347238923]
We aim to examine whether adding extra hidden layers or parameters to "blackbox ish" neural networks genuinely enhances short term price forecasting.<n>We benchmark a spectrum of models from interpretable baselines, logistic regression, XGBoost to deep architectures (DeepLOB, Conv1D+LSTM) on BTC/USDT LOB snapshots sampled at 100 ms to multi second intervals using publicly available Bybit data.
arXiv Detail & Related papers (2025-06-06T05:43:30Z) - SGD Finds then Tunes Features in Two-Layer Neural Networks with
near-Optimal Sample Complexity: A Case Study in the XOR problem [1.3597551064547502]
We consider the optimization process of minibatch descent gradient (SGD) on a 2-layer neural network with data separated by a quadratic ground truth function.
We prove that with data drawn from the $d$-dimensional Boolean hypercube labeled by the quadratic XOR'' function $y = -x_ix_j$, it is possible to train to a population error $o(1)$ with $d :textpolylog(d)$ samples.
arXiv Detail & Related papers (2023-09-26T17:57:44Z) - From Fake to Real: Pretraining on Balanced Synthetic Images to Prevent Spurious Correlations in Image Recognition [64.59093444558549]
We propose a simple, easy-to-implement, two-step training pipeline that we call From Fake to Real.
By training on real and synthetic data separately, FFR does not expose the model to the statistical differences between real and synthetic data.
Our experiments show that FFR improves worst group accuracy over the state-of-the-art by up to 20% over three datasets.
arXiv Detail & Related papers (2023-08-08T19:52:28Z) - LSTM and CNN application for core-collapse supernova search in
gravitational wave real data [0.0]
Core-collapse supernovae (CCSNe) are expected to emit gravitational wave signals that could be detected by interferometers within the Milky Way and nearby galaxies.
We show potential of machine learning (ML) for multi-label classification of different CCSNe simulated signals and noise transients using real data.
arXiv Detail & Related papers (2023-01-23T12:12:33Z) - Neural Inference of Gaussian Processes for Time Series Data of Quasars [72.79083473275742]
We introduce a new model that enables it to describe quasar spectra completely.
We also introduce a new method of inference of Gaussian process parameters, which we call $textitNeural Inference$.
The combination of both the CDRW model and Neural Inference significantly outperforms the baseline DRW and MLE.
arXiv Detail & Related papers (2022-11-17T13:01:26Z) - Two-dimensional total absorption spectroscopy with conditional
generative adversarial networks [0.22499166814992444]
We use conditional generative adversarial networks to unfold $E_x$ and $E_gamma$ data in TAS detectors.
Our model demonstrates characterization capabilities within detector resolution limits for upwards of 93% of simulated test cases.
arXiv Detail & Related papers (2022-06-23T15:57:37Z) - Deep learning for gravitational-wave data analysis: A resampling
white-box approach [62.997667081978825]
We apply Convolutional Neural Networks (CNNs) to detect gravitational wave (GW) signals of compact binary coalescences, using single-interferometer data from LIGO detectors.
CNNs were quite precise to detect noise but not sensitive enough to recall GW signals, meaning that CNNs are better for noise reduction than generation of GW triggers.
arXiv Detail & Related papers (2020-09-09T03:28:57Z) - Regularization Matters: A Nonparametric Perspective on Overparametrized
Neural Network [20.132432350255087]
Overparametrized neural networks trained by tangent descent (GD) can provably overfit any training data.
This paper studies how well overparametrized neural networks can recover the true target function in the presence of random noises.
arXiv Detail & Related papers (2020-07-06T01:02:23Z) - Quantum Algorithms for Simulating the Lattice Schwinger Model [63.18141027763459]
We give scalable, explicit digital quantum algorithms to simulate the lattice Schwinger model in both NISQ and fault-tolerant settings.
In lattice units, we find a Schwinger model on $N/2$ physical sites with coupling constant $x-1/2$ and electric field cutoff $x-1/2Lambda$.
We estimate observables which we cost in both the NISQ and fault-tolerant settings by assuming a simple target observable---the mean pair density.
arXiv Detail & Related papers (2020-02-25T19:18:36Z) - Gravitational-wave parameter estimation with autoregressive neural
network flows [0.0]
We introduce the use of autoregressive normalizing flows for rapid likelihood-free inference of binary black hole system parameters from gravitational-wave data with deep neural networks.
A normalizing flow is an invertible mapping on a sample space that can be used to induce a transformation from a simple probability distribution to a more complex one.
We build a more powerful latent variable model by incorporating autoregressive flows within the variational autoencoder framework.
arXiv Detail & Related papers (2020-02-18T15:44:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.