Estimating Multi-chirp Parameters using Curvature-guided Langevin Monte Carlo
- URL: http://arxiv.org/abs/2501.18178v1
- Date: Thu, 30 Jan 2025 07:15:38 GMT
- Title: Estimating Multi-chirp Parameters using Curvature-guided Langevin Monte Carlo
- Authors: Sattwik Basu, Debottam Dutta, Yu-Lin Wei, Romit Roy Choudhury,
- Abstract summary: This paper considers the problem of parameters from a noisy mixture of chirps.<n>We propose a modified Langevin Monte Carlo (LMC) sampler that exploits the average curvature of the objective function.
- Score: 7.832209959041259
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper considers the problem of estimating chirp parameters from a noisy mixture of chirps. While a rich body of work exists in this area, challenges remain when extending these techniques to chirps of higher order polynomials. We formulate this as a non-convex optimization problem and propose a modified Langevin Monte Carlo (LMC) sampler that exploits the average curvature of the objective function to reliably find the minimizer. Results show that our Curvature-guided LMC (CG-LMC) algorithm is robust and succeeds even in low SNR regimes, making it viable for practical applications.
Related papers
- Can Microcanonical Langevin Dynamics Leverage Mini-Batch Gradient Noise? [15.401115530780267]
Scaling inference methods such as chain Monte Carlo to high-dimensional models remains a central challenge in deep learning.<n>A promising recent proposal, microcanonical Langevin Monte Carlo, has shown state-of-the-art performance across a wide range of problems.<n>This paper addresses a fundamental question: Can microcanonical dynamics effectively leverage mini-batch gradient noise?
arXiv Detail & Related papers (2026-02-06T08:52:19Z) - Efficient Stochastic Optimisation via Sequential Monte Carlo [0.5599792629509229]
We develop sequential Monte Carlo samplers for optimisation of functions with intractable gradients.<n>Our approach replaces expensive inner sampling methods with efficient SMC approximations, which can result in significant computational gains.<n>We demonstrate the effectiveness of our approach on the reward-tuning of energy-based models within various settings.
arXiv Detail & Related papers (2026-01-29T17:13:25Z) - Quantum Speedups for Markov Chain Monte Carlo Methods with Application to Optimization [12.054017903540194]
We propose quantum algorithms that provide provable speedups for Markov Chain Monte Carlo methods.
By introducing novel techniques for gradient estimation, our algorithms improve the complexities of classical samplers.
arXiv Detail & Related papers (2025-04-04T17:44:22Z) - Fast Value Tracking for Deep Reinforcement Learning [7.648784748888187]
Reinforcement learning (RL) tackles sequential decision-making problems by creating agents that interact with their environment.
Existing algorithms often view these problem as static, focusing on point estimates for model parameters to maximize expected rewards.
Our research leverages the Kalman paradigm to introduce a novel quantification and sampling algorithm called Langevinized Kalman TemporalTD.
arXiv Detail & Related papers (2024-03-19T22:18:19Z) - A Kaczmarz-inspired approach to accelerate the optimization of neural network wavefunctions [0.7438129207086058]
We propose the Subsampled Projected Gradient-Increment Natural Descent (SPRING) to reduce this bottleneck.
SPRING combines ideas from the recently introduced minimum-step reconfiguration (MinSR) and the classical randomized Kaczmarz method for solving linear least-squares problems.
We demonstrate that SPRING outperforms both MinSR and the popular Kronecker-Factored Approximate Curvature method (KFAC) across a number of small atoms and molecules.
arXiv Detail & Related papers (2024-01-18T18:23:10Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo [104.9535542833054]
We present a scalable and effective exploration strategy based on Thompson sampling for reinforcement learning (RL)
We instead directly sample the Q function from its posterior distribution, by using Langevin Monte Carlo.
Our approach achieves better or similar results compared with state-of-the-art deep RL algorithms on several challenging exploration tasks from the Atari57 suite.
arXiv Detail & Related papers (2023-05-29T17:11:28Z) - A Model-Based Method for Minimizing CVaR and Beyond [7.751691910877239]
We develop a variant of the prox-linear method for minimizing the Conditional Value-at-Risk (CVaR) objective.
CVaR is a risk measure focused on minimizing worst-case performance, defined as the average of the top quantile of the losses.
In machine learning, such a risk measure is useful to train more robust models.
arXiv Detail & Related papers (2023-05-27T15:38:53Z) - First-Order Algorithms for Min-Max Optimization in Geodesic Metric
Spaces [93.35384756718868]
min-max algorithms have been analyzed in the Euclidean setting.
We prove that the extraiteient (RCEG) method corrected lastrate convergence at a linear rate.
arXiv Detail & Related papers (2022-06-04T18:53:44Z) - Low-variance estimation in the Plackett-Luce model via quasi-Monte Carlo
sampling [58.14878401145309]
We develop a novel approach to producing more sample-efficient estimators of expectations in the PL model.
We illustrate our findings both theoretically and empirically using real-world recommendation data from Amazon Music and the Yahoo learning-to-rank challenge.
arXiv Detail & Related papers (2022-05-12T11:15:47Z) - Parallel Stochastic Mirror Descent for MDPs [72.75921150912556]
We consider the problem of learning the optimal policy for infinite-horizon Markov decision processes (MDPs)
Some variant of Mirror Descent is proposed for convex programming problems with Lipschitz-continuous functionals.
We analyze this algorithm in a general case and obtain an estimate of the convergence rate that does not accumulate errors during the operation of the method.
arXiv Detail & Related papers (2021-02-27T19:28:39Z) - Marginalised Gaussian Processes with Nested Sampling [10.495114898741203]
Gaussian Process (GPs) models are a rich distribution over functions with inductive biases controlled by a kernel function.
This work presents an alternative learning procedure where the hyperparameters of the kernel function are marginalised using Nested Sampling (NS)
arXiv Detail & Related papers (2020-10-30T16:04:35Z) - Stochastic Optimization with Heavy-Tailed Noise via Accelerated Gradient
Clipping [69.9674326582747]
We propose a new accelerated first-order method called clipped-SSTM for smooth convex optimization with heavy-tailed distributed noise in gradients.
We prove new complexity that outperform state-of-the-art results in this case.
We derive the first non-trivial high-probability complexity bounds for SGD with clipping without light-tails assumption on the noise.
arXiv Detail & Related papers (2020-05-21T17:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.