Simultaneous Optimization of Geodesics and Fréchet Means
- URL: http://arxiv.org/abs/2511.04301v1
- Date: Thu, 06 Nov 2025 12:08:15 GMT
- Title: Simultaneous Optimization of Geodesics and Fréchet Means
- Authors: Frederik Möbius Rygaard, Søren Hauberg, Steen Markvorsen,
- Abstract summary: A central part of geometric statistics is to compute the Fr'echet mean.<n>We introduce the GEORCE-FM algorithm to simultaneously compute the Fr'echet mean and Riemannian distances in each iteration in a local chart.<n>We show that GEORCE-FM has global convergence and local quadratic convergence and prove that the adaptive extension converges in expectation to the Fr'echet mean.
- Score: 11.807982858405056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A central part of geometric statistics is to compute the Fr\'echet mean. This is a well-known intrinsic mean on a Riemannian manifold that minimizes the sum of squared Riemannian distances from the mean point to all other data points. The Fr\'echet mean is simple to define and generalizes the Euclidean mean, but for most manifolds even minimizing the Riemannian distance involves solving an optimization problem. Therefore, numerical computations of the Fr\'echet mean require solving an embedded optimization problem in each iteration. We introduce the GEORCE-FM algorithm to simultaneously compute the Fr\'echet mean and Riemannian distances in each iteration in a local chart, making it faster than previous methods. We extend the algorithm to Finsler manifolds and introduce an adaptive extension such that GEORCE-FM scales to a large number of data points. Theoretically, we show that GEORCE-FM has global convergence and local quadratic convergence and prove that the adaptive extension converges in expectation to the Fr\'echet mean. We further empirically demonstrate that GEORCE-FM outperforms existing baseline methods to estimate the Fr\'echet mean in terms of both accuracy and runtime.
Related papers
- Riemannian stochastic optimization methods avoid strict saddle points [68.80251170757647]
We show that policies under study avoid strict saddle points / submanifolds with probability 1.
This result provides an important sanity check as it shows that, almost always, the limit state of an algorithm can only be a local minimizer.
arXiv Detail & Related papers (2023-11-04T11:12:24Z) - Decentralized Riemannian Conjugate Gradient Method on the Stiefel
Manifold [59.73080197971106]
This paper presents a first-order conjugate optimization method that converges faster than the steepest descent method.
It aims to achieve global convergence over the Stiefel manifold.
arXiv Detail & Related papers (2023-08-21T08:02:16Z) - Exploiting Structure for Optimal Multi-Agent Bayesian Decentralized
Estimation [4.320393382724066]
Key challenge in Bayesian decentralized data fusion is the rumor propagation' or double counting' phenomenon.
We show that by exploiting the probabilistic independence structure in multi-agent decentralized fusion problems a tighter bound can be found.
We then test our new non-monolithic CI algorithm on a large-scale target tracking simulation and show that it achieves a tighter bound and a more accurate estimate.
arXiv Detail & Related papers (2023-07-20T05:16:33Z) - Curvature-Independent Last-Iterate Convergence for Games on Riemannian
Manifolds [77.4346324549323]
We show that a step size agnostic to the curvature of the manifold achieves a curvature-independent and linear last-iterate convergence rate.
To the best of our knowledge, the possibility of curvature-independent rates and/or last-iterate convergence has not been considered before.
arXiv Detail & Related papers (2023-06-29T01:20:44Z) - Decentralized Riemannian natural gradient methods with Kronecker-product
approximations [11.263837420265594]
We present an efficient decentralized natural gradient descent (DRNGD) method for solving decentralized manifold optimization problems.
By performing the communications over the Kronecker factors, a high-quality approximation of the RFIM can be obtained in a low cost.
arXiv Detail & Related papers (2023-03-16T19:36:31Z) - Flow Matching on General Geometries [43.252817099263744]
We propose a simple yet powerful framework for training continuous normalizing flows on manifold geometries.
We show that it is simulation-free on simple geometries, does not require divergence, and computes its target vector field in closed-form.
Our method achieves state-of-the-art performance on many real-world non-Euclidean datasets.
arXiv Detail & Related papers (2023-02-07T18:21:24Z) - The Dynamics of Riemannian Robbins-Monro Algorithms [101.29301565229265]
We propose a family of Riemannian algorithms generalizing and extending the seminal approximation framework of Robbins and Monro.
Compared to their Euclidean counterparts, Riemannian algorithms are much less understood due to lack of a global linear structure on the manifold.
We provide a general template of almost sure convergence results that mirrors and extends the existing theory for Euclidean Robbins-Monro schemes.
arXiv Detail & Related papers (2022-06-14T12:30:11Z) - First-Order Algorithms for Min-Max Optimization in Geodesic Metric
Spaces [93.35384756718868]
min-max algorithms have been analyzed in the Euclidean setting.
We prove that the extraiteient (RCEG) method corrected lastrate convergence at a linear rate.
arXiv Detail & Related papers (2022-06-04T18:53:44Z) - Averaging on the Bures-Wasserstein manifold: dimension-free convergence
of gradient descent [15.136397170510834]
We prove new geodesic convexity results which provide stronger control of the iterates, a free convergence.
Our techniques also enable the analysis of two related notions of averaging, the entropically-regularized barycenter and the geometric median.
arXiv Detail & Related papers (2021-06-16T01:05:19Z) - Bayesian Quadrature on Riemannian Data Manifolds [79.71142807798284]
A principled way to model nonlinear geometric structure inherent in data is provided.
However, these operations are typically computationally demanding.
In particular, we focus on Bayesian quadrature (BQ) to numerically compute integrals over normal laws.
We show that by leveraging both prior knowledge and an active exploration scheme, BQ significantly reduces the number of required evaluations.
arXiv Detail & Related papers (2021-02-12T17:38:04Z) - Riemannian Stochastic Proximal Gradient Methods for Nonsmooth
Optimization over the Stiefel Manifold [7.257751371276488]
R-ProxSGD and R-ProxSPB are generalizations of proximal SGD and proximal SpiderBoost.
R-ProxSPB algorithm finds an $epsilon$-stationary point with $O(epsilon-3)$ IFOs in the online case, and $O(n+sqrtnepsilon-3)$ IFOs in the finite-sum case.
arXiv Detail & Related papers (2020-05-03T23:41:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.