Optimal Stopping in Latent Diffusion Models
- URL: http://arxiv.org/abs/2510.08409v1
- Date: Thu, 09 Oct 2025 16:28:48 GMT
- Title: Optimal Stopping in Latent Diffusion Models
- Authors: Yu-Han Wu, Quentin Berthet, Gérard Biau, Claire Boyer, Romuald Elie, Pierre Marion,
- Abstract summary: We identify and analyze a surprising phenomenon of Latent Diffusion Models (LDMs) where the final steps of the diffusion can degrade sample quality.<n>We provide a principled explanation by analyzing the interaction between latent dimension and stopping time.
- Score: 22.471547966218278
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We identify and analyze a surprising phenomenon of Latent Diffusion Models (LDMs) where the final steps of the diffusion can degrade sample quality. In contrast to conventional arguments that justify early stopping for numerical stability, this phenomenon is intrinsic to the dimensionality reduction in LDMs. We provide a principled explanation by analyzing the interaction between latent dimension and stopping time. Under a Gaussian framework with linear autoencoders, we characterize the conditions under which early stopping is needed to minimize the distance between generated and target distributions. More precisely, we show that lower-dimensional representations benefit from earlier termination, whereas higher-dimensional latent spaces require later stopping time. We further establish that the latent dimension interplays with other hyperparameters of the problem such as constraints in the parameters of score matching. Experiments on synthetic and real datasets illustrate these properties, underlining that early stopping can improve generative quality. Together, our results offer a theoretical foundation for understanding how the latent dimension influences the sample quality, and highlight stopping time as a key hyperparameter in LDMs.
Related papers
- Multi-Parameter Multi-Critical Metrology of the Dicke Model [10.440724472122662]
This work exploits the hypersensitivity of quantum systems near phase transitions to achieve enhanced precision in parameter estimation.<n>We show that multi parameter estimation is feasible but can also retain divergent precision scaling.<n>Our results pave the way for practical quantum sensors operating near phase transitions.
arXiv Detail & Related papers (2026-03-03T19:06:55Z) - Dispersive Hong-Ou-Mandel Interference with Finite Coincidence Windows [0.0]
Hong-Ou-Mandel (HOM) interference is a fundamental tool for assessing photon indistinguishability in quantum information processing.<n>Modern time-tagging modules, which effectively acts as a temporal filter, break the standard dispersion cancellation condition.<n>We derive an analytical model for type-II SPDC processes that predicts a modification of the HOM dip shape.
arXiv Detail & Related papers (2026-02-20T11:44:25Z) - From Observations to States: Latent Time Series Forecasting [65.98504021691666]
We propose Latent Time Series Forecasting (LatentTSF), a novel paradigm that shifts TSF from observation regression to latent state prediction.<n>Specifically, LatentTSF employs an AutoEncoder to project observations at each time step into a higher-dimensional latent state space.<n>Our proposed latent objectives implicitly maximize mutual information between predicted latent states and ground-truth states and observations.
arXiv Detail & Related papers (2026-01-30T20:39:44Z) - Particle Dynamics for Latent-Variable Energy-Based Models [12.84928511163926]
Latent-variable energy-based models (LVEBMs) assign a single normalized energy to joint pairs of observed data and latent variables.<n>We recast maximum-likelihood training as a saddle problem over distributions on the latent and joint gradients.<n>We prove existence and convergence under standard smoothness and dissipativity assumptions, with decay rates in KL divergence and Wasserstein-2 distance.
arXiv Detail & Related papers (2025-10-17T09:04:49Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.<n>We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.<n>Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - On latent dynamics learning in nonlinear reduced order modeling [0.6249768559720122]
We present the novel mathematical framework of latent dynamics models (LDMs) for reduced order modeling of parameterized nonlinear time-dependent PDEs.<n>A time-continuous setting is employed to derive error and stability estimates for the LDM approximation of the full order model (FOM) solution.<n>Deep neural networks approximate the discrete LDM components, while providing a bounded approximation error with respect to the FOM.
arXiv Detail & Related papers (2024-08-27T16:35:06Z) - A Study of Posterior Stability for Time-Series Latent Diffusion [59.41969496514184]
We first show that posterior collapse will reduce latent diffusion to a variational autoencoder (VAE), making it less expressive.
We then introduce a principled method: dependency measure, that quantifies the sensitivity of a recurrent decoder to input variables.
Building on our theoretical and empirical studies, we introduce a new framework that extends latent diffusion and has a stable posterior.
arXiv Detail & Related papers (2024-05-22T21:54:12Z) - Multi-fidelity reduced-order surrogate modeling [5.346062841242067]
We present a new data-driven strategy that combines dimensionality reduction with multi-fidelity neural network surrogates.
We show that the onset of instabilities and transients are well captured by this surrogate technique.
arXiv Detail & Related papers (2023-09-01T08:16:53Z) - Convergence of mean-field Langevin dynamics: Time and space
discretization, stochastic gradient, and variance reduction [49.66486092259376]
The mean-field Langevin dynamics (MFLD) is a nonlinear generalization of the Langevin dynamics that incorporates a distribution-dependent drift.
Recent works have shown that MFLD globally minimizes an entropy-regularized convex functional in the space of measures.
We provide a framework to prove a uniform-in-time propagation of chaos for MFLD that takes into account the errors due to finite-particle approximation, time-discretization, and gradient approximation.
arXiv Detail & Related papers (2023-06-12T16:28:11Z) - Switching Autoregressive Low-rank Tensor Models [12.461139675114818]
We show how to switch autoregressive low-rank tensor (SALT) models.
SALT parameterizes the tensor of an ARHMM with a low-rank factorization to control the number of parameters.
We prove theoretical and discuss practical connections between SALT, linear dynamical systems, and SLDSs.
arXiv Detail & Related papers (2023-06-05T22:25:28Z) - Reconstructing Graph Diffusion History from a Single Snapshot [87.20550495678907]
We propose a novel barycenter formulation for reconstructing Diffusion history from A single SnapsHot (DASH)
We prove that estimation error of diffusion parameters is unavoidable due to NP-hardness of diffusion parameter estimation.
We also develop an effective solver named DIffusion hiTting Times with Optimal proposal (DITTO)
arXiv Detail & Related papers (2023-06-01T09:39:32Z) - Supporting Optimal Phase Space Reconstructions Using Neural Network
Architecture for Time Series Modeling [68.8204255655161]
We propose an artificial neural network with a mechanism to implicitly learn the phase spaces properties.
Our approach is either as competitive as or better than most state-of-the-art strategies.
arXiv Detail & Related papers (2020-06-19T21:04:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.