Bregman-Hausdorff divergence: strengthening the connections between computational geometry and machine learning
- URL: http://arxiv.org/abs/2504.07322v1
- Date: Wed, 09 Apr 2025 22:42:29 GMT
- Title: Bregman-Hausdorff divergence: strengthening the connections between computational geometry and machine learning
- Authors: Tuyen Pham, Hana Dal Poz Kouřimská, Hubert Wagner,
- Abstract summary: We focus on the family of Bregman divergences, which includes the popular Kullback--Leibler divergence.<n>As a proof of concept, we use the resulting Bregman--Hausdorff divergence to compare two collections of probabilistic predictions.<n>The algorithms we propose are surprisingly efficient even for large inputs with hundreds of dimensions.
- Score: 0.6554326244334868
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The purpose of this paper is twofold. On a technical side, we propose an extension of the Hausdorff distance from metric spaces to spaces equipped with asymmetric distance measures. Specifically, we focus on the family of Bregman divergences, which includes the popular Kullback--Leibler divergence (also known as relative entropy). As a proof of concept, we use the resulting Bregman--Hausdorff divergence to compare two collections of probabilistic predictions produced by different machine learning models trained using the relative entropy loss. The algorithms we propose are surprisingly efficient even for large inputs with hundreds of dimensions. In addition to the introduction of this technical concept, we provide a survey. It outlines the basics of Bregman geometry, as well as computational geometry algorithms. We focus on algorithms that are compatible with this geometry and are relevant for machine learning.
Related papers
- CayleyPy RL: Pathfinding and Reinforcement Learning on Cayley Graphs [0.0]
This paper is the second in a series of studies on developing efficient artificial intelligence-based approaches to pathfinding on large graphs.
We present a novel combination of a reinforcement learning approach with a more direct diffusion distance approach from the first paper.
We provide strong support for the OEIS-A186783 conjecture that the diameter is equal to n(n-1)/2 by machine learning and mathematical methods.
arXiv Detail & Related papers (2025-02-25T21:53:41Z) - The Fisher-Rao geometry of CES distributions [50.50897590847961]
The Fisher-Rao information geometry allows for leveraging tools from differential geometry.
We will present some practical uses of these geometric tools in the framework of elliptical distributions.
arXiv Detail & Related papers (2023-10-02T09:23:32Z) - Neural Latent Geometry Search: Product Manifold Inference via
Gromov-Hausdorff-Informed Bayesian Optimization [21.97865037637575]
We mathematically define this novel formulation and coin it as neural latent geometry search (NLGS)
We propose a novel notion of distance between candidate latent geometries based on the Gromov-Hausdorff distance from metric geometry.
We then design a graph search space based on the notion of smoothness between latent geometries and employ the calculated as an additional inductive bias.
arXiv Detail & Related papers (2023-09-09T14:29:22Z) - Gromov-Hausdorff Distances for Comparing Product Manifolds of Model
Spaces [21.97865037637575]
We introduce a novel notion of distance between candidate latent geometries using the Gromov-Hausdorff distance from metric geometry.
We propose using a graph search space that uses the estimated Gromov-Hausdorff distances to search for the optimal latent geometry.
arXiv Detail & Related papers (2023-09-09T11:17:06Z) - On a class of geodesically convex optimization problems solved via
Euclidean MM methods [50.428784381385164]
We show how a difference of Euclidean convexization functions can be written as a difference of different types of problems in statistics and machine learning.
Ultimately, we helps the broader broader the broader the broader the broader the work.
arXiv Detail & Related papers (2022-06-22T23:57:40Z) - Neural Bregman Divergences for Distance Learning [60.375385370556145]
We propose a new approach to learning arbitrary Bregman divergences in a differentiable manner via input convex neural networks.
We show that our method more faithfully learns divergences over a set of both new and previously studied tasks.
Our tests further extend to known asymmetric, but non-Bregman tasks, where our method still performs competitively despite misspecification.
arXiv Detail & Related papers (2022-06-09T20:53:15Z) - Group invariant machine learning by fundamental domain projections [0.5156484100374058]
We approach the well-studied problem of supervised group invariant and equivariant machine learning from the point of view of geometric topology.
We propose a novel approach using a pre-processing step, which involves projecting the input data into a geometric space.
This new data can then be the input for an arbitrary machine learning model.
arXiv Detail & Related papers (2022-02-04T14:45:57Z) - A singular Riemannian geometry approach to Deep Neural Networks II.
Reconstruction of 1-D equivalence classes [78.120734120667]
We build the preimage of a point in the output manifold in the input space.
We focus for simplicity on the case of neural networks maps from n-dimensional real spaces to (n - 1)-dimensional real spaces.
arXiv Detail & Related papers (2021-12-17T11:47:45Z) - Neural Operator: Graph Kernel Network for Partial Differential Equations [57.90284928158383]
This work is to generalize neural networks so that they can learn mappings between infinite-dimensional spaces (operators)
We formulate approximation of the infinite-dimensional mapping by composing nonlinear activation functions and a class of integral operators.
Experiments confirm that the proposed graph kernel network does have the desired properties and show competitive performance compared to the state of the art solvers.
arXiv Detail & Related papers (2020-03-07T01:56:20Z) - Fast and Robust Comparison of Probability Measures in Heterogeneous
Spaces [62.35667646858558]
We introduce the Anchor Energy (AE) and Anchor Wasserstein (AW) distances, which are respectively the energy and Wasserstein distances instantiated on such representations.
Our main contribution is to propose a sweep line algorithm to compute AE emphexactly in log-quadratic time, where a naive implementation would be cubic.
We show that AE and AW perform well in various experimental settings at a fraction of the computational cost of popular GW approximations.
arXiv Detail & Related papers (2020-02-05T03:09:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.