Long Range Frequency Tuning for QML
- URL: http://arxiv.org/abs/2602.23409v1
- Date: Thu, 26 Feb 2026 14:56:37 GMT
- Title: Long Range Frequency Tuning for QML
- Authors: Michael Poppel, Jonas Stein, Sebastian Wölckert, Markus Baumann, Claudia Linnhoff-Popien,
- Abstract summary: Trainable-frequency approaches theoretically reduce this to match the target spectrum size, requiring only as many encoding gates as frequencies in the target spectrum.<n>We show that gradient-based optimization can drive prefactors to arbitrary target values, but they exhibit limited trainability.<n>We propose grid-based initialization using ternary encodings, which generate dense integer frequency spectra.
- Score: 2.34764060945577
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Quantum machine learning models using angle encoding naturally represent truncated Fourier series, providing universal function approximation capabilities with sufficient circuit depth. For unary fixed-frequency encodings, circuit depth scales as O(omega_max * (omega_max + epsilon^{-2})) with target frequency magnitude omega_max and precision epsilon. Trainable-frequency approaches theoretically reduce this to match the target spectrum size, requiring only as many encoding gates as frequencies in the target spectrum. Despite this compelling efficiency, their practical effectiveness hinges on a key assumption: that gradient-based optimization can drive prefactors to arbitrary target values. We demonstrate through systematic experiments that frequency prefactors exhibit limited trainability: movement is constrained to approximately +/-1 units with typical learning rates. When target frequencies lie outside this reachable range, optimization frequently fails. To overcome this frequency reachability limitation, we propose grid-based initialization using ternary encodings, which generate dense integer frequency spectra. While this approach requires O(log_3(omega_max)) encoding gates -- more than the theoretical optimum but exponentially fewer than fixed-frequency methods -- it ensures target frequencies lie within the locally reachable range. On synthetic targets with three shifted high frequencies, ternary grid initialization achieves a median R^2 score of 0.9969, compared to 0.1841 for the trainable-frequency baseline. For the real-world Flight Passengers dataset, ternary grid initialization achieves a median R^2 score of 0.9671, representing a 22.8% improvement over trainable-frequency initialization (median R^2 = 0.7876).
Related papers
- Fast frequency reconstruction using Deep Learning for event recognition in ring laser data [0.0]
We present a neural network approach capable of reconstructing frequencies of several hundred Hertz within approximately 10 milliseconds.<n>The method outperforms standard Fourier-based techniques, improving frequency estimation precision by a factor of 2 in the operational range of GINGERINO.<n>In addition to fast frequency estimation, we introduce an automated classification framework to identify physical disturbances in the signal, such as laser instabilities and seismic events.
arXiv Detail & Related papers (2025-10-01T15:18:31Z) - Mitigating Exponential Mixed Frequency Growth through Frequency Selection [3.243345719637229]
Angle encoding has emerged as a popular choice as feature map (FM) for embedding classical data into quantum models.<n>We show that training failures can occur even when all relevant frequencies are theoretically accessible.<n>We introduce frequency selection as a practical solution that reduces parameter requirements.
arXiv Detail & Related papers (2025-08-14T11:10:07Z) - Wideband RF Radiance Field Modeling Using Frequency-embedded 3D Gaussian Splatting [28.147938573798367]
We present an innovative frequency-embedded 3D Gaussian splatting (3DGS) algorithm for wideband radio-frequency (RF) radiance field modeling.<n>We propose a large-scale power angular spectrum (PAS) dataset containing 50000 samples ranging from 1 to 100 GHz in 6 indoor environments.<n>Our approach achieves an average Structural Similarity Index Measure (SSIM) up to 0.72, and a significant improvement up to 17.8% compared to the current state-of-the-art (SOTA) methods.
arXiv Detail & Related papers (2025-05-27T04:48:26Z) - Chirp Localization via Fine-Tuned Transformer Model: A Proof-of-Concept Study [0.23020018305241333]
Chirp-like patterns in EEG spectrograms are key biomarkers for seizure dynamics.<n>This study bridges this gap by fine-tuning a Vision Transformer (ViT) model and Low-Rank Adaptation (LoRA)<n>We generated 100000 spectrograms with chirp parameters, creating the first large-scale benchmark for chirp localization.
arXiv Detail & Related papers (2025-03-24T14:27:07Z) - Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning [81.98675881423131]
This research addresses the challenge of developing a universal deepfake detector that can effectively identify unseen deepfake images.
Existing frequency-based paradigms have relied on frequency-level artifacts introduced during the up-sampling in GAN pipelines to detect forgeries.
We introduce a novel frequency-aware approach called FreqNet, centered around frequency domain learning, specifically designed to enhance the generalizability of deepfake detectors.
arXiv Detail & Related papers (2024-03-12T01:28:00Z) - Transform Once: Efficient Operator Learning in Frequency Domain [69.74509540521397]
We study deep neural networks designed to harness the structure in frequency domain for efficient learning of long-range correlations in space or time.
This work introduces a blueprint for frequency domain learning through a single transform: transform once (T1)
arXiv Detail & Related papers (2022-11-26T01:56:05Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - FastFlowNet: A Lightweight Network for Fast Optical Flow Estimation [81.76975488010213]
Dense optical flow estimation plays a key role in many robotic vision tasks.
Current networks often occupy large number of parameters and require heavy computation costs.
Our proposed FastFlowNet works in the well-known coarse-to-fine manner with following innovations.
arXiv Detail & Related papers (2021-03-08T03:09:37Z) - Learning Frequency Domain Approximation for Binary Neural Networks [68.79904499480025]
We propose to estimate the gradient of sign function in the Fourier frequency domain using the combination of sine functions for training BNNs.
The experiments on several benchmark datasets and neural architectures illustrate that the binary network learned using our method achieves the state-of-the-art accuracy.
arXiv Detail & Related papers (2021-03-01T08:25:26Z) - The cost of universality: A comparative study of the overhead of state
distillation and code switching with color codes [63.62764375279861]
We compare two leading FT implementations of the T gate in 2D color codes under circuit noise.
We find a circuit noise threshold of 0.07(1)% for the T gate via code switching, almost an order of magnitude below that achievable by state distillation in the same setting.
arXiv Detail & Related papers (2021-01-06T19:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.