CurveFlow: Curvature-Guided Flow Matching for Image Generation
- URL: http://arxiv.org/abs/2508.15093v2
- Date: Sun, 24 Aug 2025 04:41:55 GMT
- Title: CurveFlow: Curvature-Guided Flow Matching for Image Generation
- Authors: Yan Luo, Drake Du, Hao Huang, Yi Fang, Mengyu Wang,
- Abstract summary: Existing rectified flow models are based on linear trajectories between data and noise distributions.<n>This linearity enforces zero curvature, which can inadvertently force the image generation process through low-probability regions of the data manifold.<n>We introduce CurveFlow, a novel flow matching framework designed to learn smooth, non-linear trajectories by incorporating curvature guidance into the flow path.
- Score: 11.836900973675297
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Existing rectified flow models are based on linear trajectories between data and noise distributions. This linearity enforces zero curvature, which can inadvertently force the image generation process through low-probability regions of the data manifold. A key question remains underexplored: how does the curvature of these trajectories correlate with the semantic alignment between generated images and their corresponding captions, i.e., instructional compliance? To address this, we introduce CurveFlow, a novel flow matching framework designed to learn smooth, non-linear trajectories by directly incorporating curvature guidance into the flow path. Our method features a robust curvature regularization technique that penalizes abrupt changes in the trajectory's intrinsic dynamics.Extensive experiments on MS COCO 2014 and 2017 demonstrate that CurveFlow achieves state-of-the-art performance in text-to-image generation, significantly outperforming both standard rectified flow variants and other non-linear baselines like Rectified Diffusion. The improvements are especially evident in semantic consistency metrics such as BLEU, METEOR, ROUGE, and CLAIR. This confirms that our curvature-aware modeling substantially enhances the model's ability to faithfully follow complex instructions while simultaneously maintaining high image quality. The code is made publicly available at https://github.com/Harvard-AI-and-Robotics-Lab/CurveFlow.
Related papers
- ArcFlow: Unleashing 2-Step Text-to-Image Generation via High-Precision Non-Linear Flow Distillation [87.54456066636811]
Diffusion models suffer from significant inference cost due to their reliance on sequential denoising steps.<n>ArcFlow is a few-step distillation framework that explicitly employs non-linear flow trajectories to approximate pre-trained teacher trajectories.<n>It achieves a 40x speedup with 2 NFEs over the original multi-step teachers without significant quality degradation.
arXiv Detail & Related papers (2026-02-09T18:56:14Z) - SplineFlow: Flow Matching for Dynamical Systems with B-Spline Interpolants [14.711575625163045]
SplineFlow is a theoretically grounded flow matching algorithm that jointly models conditional paths across observations via B-spline.<n>We show how SplineFlow exploits the smoothness and stability of B-spline bases learn the complex underlying dynamics while ensuring the multi-marginal requirements are met.
arXiv Detail & Related papers (2026-01-30T15:19:48Z) - FlowPath: Learning Data-Driven Manifolds with Invertible Flows for Robust Irregularly-sampled Time Series Classification [14.643457217551484]
We propose FlowPath, a novel approach that learns the geometry of the control path via an invertible neural flow.<n>We show that FlowPath consistently achieves statistically significant improvements in classification accuracy over baselines using fixed interpolants or non-invertible architectures.
arXiv Detail & Related papers (2025-11-13T22:59:26Z) - Rectified-CFG++ for Flow Based Models [26.896426878221718]
We present Rectified-C++, an adaptive predictor-corrector guidance that couples the deterministic efficiency of rectified flows with a geometry-aware conditioning rule.<n>Experiments on large-scale text-to-image models (Flux, Stable Diffusion 3/3.5, Lumina) show that Rectified-C++ consistently outperforms standard CFG on benchmark datasets.
arXiv Detail & Related papers (2025-10-09T00:00:47Z) - Bures-Wasserstein Flow Matching for Graph Generation [11.634569150774979]
Graph generation has emerged as a critical task in fields ranging from molecule design to drug discovery.<n>We introduce BWFlow, a flow-matching framework for graph generation that respects the underlying geometry of graphs.<n>The novel framework can be adapted to both continuous and discrete flow-matching algorithms.
arXiv Detail & Related papers (2025-06-16T21:36:56Z) - Solving Inverse Problems with FLAIR [59.02385492199431]
Flow-based latent generative models are able to generate images with remarkable quality, even enabling text-to-image generation.<n>We present FLAIR, a novel training free variational framework that leverages flow-based generative models as a prior for inverse problems.<n>Results on standard imaging benchmarks demonstrate that FLAIR consistently outperforms existing diffusion- and flow-based methods in terms of reconstruction quality and sample diversity.
arXiv Detail & Related papers (2025-06-03T09:29:47Z) - FlowAlign: Trajectory-Regularized, Inversion-Free Flow-based Image Editing [47.908940130654535]
FlowAlign is an inversion-free flow-based framework for consistent image editing with optimal control-based trajectory control.<n>Our terminal point regularization is shown to balance semantic alignment with the edit prompt and structural consistency with the source image along the trajectory.<n>FlowAlign outperforms existing methods in both source preservation and editing controllability.
arXiv Detail & Related papers (2025-05-29T06:33:16Z) - Steering Rectified Flow Models in the Vector Field for Controlled Image Generation [53.965218831845995]
Diffusion models (DMs) excel in photorealism, image editing, and solving inverse problems, aided by classifier-free guidance and image inversion techniques.<n>Existing DM-based methods often require additional training, lack generalization to pretrained latent models, underperform, and demand significant computational resources due to extensive backpropagation through ODE solvers and inversion processes.<n>We propose FlowChef, which leverages the vector field to steer the denoising trajectory for controlled image generation tasks, facilitated by gradient skipping.<n>FlowChef significantly outperforms baselines in terms of performance, memory, and time requirements, achieving new state-of-the
arXiv Detail & Related papers (2024-11-27T19:04:40Z) - FlowIE: Efficient Image Enhancement via Rectified Flow [71.6345505427213]
FlowIE is a flow-based framework that estimates straight-line paths from an elementary distribution to high-quality images.
Our contributions are rigorously validated through comprehensive experiments on synthetic and real-world datasets.
arXiv Detail & Related papers (2024-06-01T17:29:29Z) - D-Flow: Differentiating through Flows for Controlled Generation [37.80603174399585]
We introduce D-Flow, a framework for controlling the generation process by differentiating through the flow.
We motivate this framework by our key observation stating that for Diffusion/FM models trained with Gaussian probability paths, differentiating through the generation process projects gradient on the data manifold.
We validate our framework on linear and non-linear controlled generation problems including: image and audio inverse problems and conditional molecule generation reaching state of the art performance across all.
arXiv Detail & Related papers (2024-02-21T18:56:03Z) - Guided Flows for Generative Modeling and Decision Making [55.42634941614435]
We show that Guided Flows significantly improves the sample quality in conditional image generation and zero-shot text synthesis-to-speech.
Notably, we are first to apply flow models for plan generation in the offline reinforcement learning setting ax speedup in compared to diffusion models.
arXiv Detail & Related papers (2023-11-22T15:07:59Z) - GAFlow: Incorporating Gaussian Attention into Optical Flow [62.646389181507764]
We push Gaussian Attention (GA) into the optical flow models to accentuate local properties during representation learning.
We introduce a novel Gaussian-Constrained Layer (GCL) which can be easily plugged into existing Transformer blocks.
For reliable motion analysis, we provide a new Gaussian-Guided Attention Module (GGAM)
arXiv Detail & Related papers (2023-09-28T07:46:01Z) - Flow Straight and Fast: Learning to Generate and Transfer Data with
Rectified Flow [32.459587479351846]
We present rectified flow, a surprisingly simple approach to learning (neural) ordinary differential equation (ODE) models.
We show that rectified flow performs superbly on image generation, image-to-image translation, and domain adaptation.
arXiv Detail & Related papers (2022-09-07T08:59:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.