Fast hyperboloid decision tree algorithms
- URL: http://arxiv.org/abs/2310.13841v2
- Date: Mon, 4 Mar 2024 20:02:03 GMT
- Title: Fast hyperboloid decision tree algorithms
- Authors: Philippe Chlenski, Ethan Turok, Antonio Moretti, Itsik Pe'er
- Abstract summary: We present hyperDT, a novel extension of decision tree algorithms into hyperbolic space.
Our approach is conceptually straightforward and maintains constant-time decision complexity.
Building upon hyperDT we introduce hyperRF, a hyperbolic random forest model.
- Score: 0.6656737591902598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hyperbolic geometry is gaining traction in machine learning for its
effectiveness at capturing hierarchical structures in real-world data.
Hyperbolic spaces, where neighborhoods grow exponentially, offer substantial
advantages and consistently deliver state-of-the-art results across diverse
applications. However, hyperbolic classifiers often grapple with computational
challenges. Methods reliant on Riemannian optimization frequently exhibit
sluggishness, stemming from the increased computational demands of operations
on Riemannian manifolds. In response to these challenges, we present hyperDT, a
novel extension of decision tree algorithms into hyperbolic space. Crucially,
hyperDT eliminates the need for computationally intensive Riemannian
optimization, numerically unstable exponential and logarithmic maps, or
pairwise comparisons between points by leveraging inner products to adapt
Euclidean decision tree algorithms to hyperbolic space. Our approach is
conceptually straightforward and maintains constant-time decision complexity
while mitigating the scalability issues inherent in high-dimensional Euclidean
spaces. Building upon hyperDT we introduce hyperRF, a hyperbolic random forest
model. Extensive benchmarking across diverse datasets underscores the superior
performance of these models, providing a swift, precise, accurate, and
user-friendly toolkit for hyperbolic data analysis.
Related papers
- Hyperbolic Fine-tuning for Large Language Models [56.54715487997674]
This study investigates the non-Euclidean characteristics of large language models (LLMs)
We show that token embeddings exhibit a high degree of hyperbolicity, indicating a latent tree-like structure in the embedding space.
We introduce a new method called hyperbolic low-rank efficient fine-tuning, HypLoRA, that performs low-rank adaptation directly on the hyperbolic manifold.
arXiv Detail & Related papers (2024-10-05T02:58:25Z) - From Semantics to Hierarchy: A Hybrid Euclidean-Tangent-Hyperbolic Space Model for Temporal Knowledge Graph Reasoning [1.1372536310854844]
Temporal knowledge graph (TKG) reasoning predicts future events based on historical data.
Existing Euclidean models excel at capturing semantics but struggle with hierarchy.
We propose a novel hybrid geometric space approach that leverages the strengths of both Euclidean and hyperbolic models.
arXiv Detail & Related papers (2024-08-30T10:33:08Z) - Discovering symbolic expressions with parallelized tree search [59.92040079807524]
Symbolic regression plays a crucial role in scientific research thanks to its capability of discovering concise and interpretable mathematical expressions from data.
Existing algorithms have faced a critical bottleneck of accuracy and efficiency over a decade when handling problems of complexity.
We introduce a parallelized tree search (PTS) model to efficiently distill generic mathematical expressions from limited data.
arXiv Detail & Related papers (2024-07-05T10:41:15Z) - Convex Relaxation for Solving Large-Margin Classifiers in Hyperbolic Space [29.564717407207528]
Hyperbolic spaces have been recognized for their outstanding performance in handling data.
Previous attempts to solve this problem using a gradient descent approach have failed.
In this paper we show how we can effectively approximate the optima using sparse moment of relaxations.
arXiv Detail & Related papers (2024-05-27T14:19:53Z) - Hyperbolic Delaunay Geometric Alignment [52.835250875177756]
We propose a similarity score for comparing datasets in a hyperbolic space.
The core idea is counting the edges of the hyperbolic Delaunay graph connecting datapoints across the given sets.
We provide an empirical investigation on synthetic and real-life biological data and demonstrate that HyperDGA outperforms the hyperbolic version of classical distances between sets.
arXiv Detail & Related papers (2024-04-12T17:14:58Z) - Accelerating hyperbolic t-SNE [7.411478341945197]
This paper introduces the first acceleration structure for hyperbolic embeddings, building upon a polar quadtree.
We demonstrate that it computes embeddings of similar quality in significantly less time.
arXiv Detail & Related papers (2024-01-23T12:59:40Z) - Scaling Riemannian Diffusion Models [68.52820280448991]
We show that our method enables us to scale to high dimensional tasks on nontrivial manifold.
We model QCD densities on $SU(n)$ lattices and contrastively learned embeddings on high dimensional hyperspheres.
arXiv Detail & Related papers (2023-10-30T21:27:53Z) - HRCF: Enhancing Collaborative Filtering via Hyperbolic Geometric
Regularization [52.369435664689995]
We introduce a textitHyperbolic Regularization powered Collaborative Filtering (HRCF) and design a geometric-aware hyperbolic regularizer.
Specifically, the proposal boosts optimization procedure via the root alignment and origin-aware penalty.
Our proposal is able to tackle the over-smoothing problem caused by hyperbolic aggregation and also brings the models a better discriminative ability.
arXiv Detail & Related papers (2022-04-18T06:11:44Z) - Provably Accurate and Scalable Linear Classifiers in Hyperbolic Spaces [39.71927912296049]
We propose a unified framework for learning scalable and simple hyperbolic linear classifiers.
The gist of our approach is to focus on Poincar'e ball models and formulate the classification problems using tangent space formalisms.
The excellent performance of the Poincar'e second-order and strategic perceptrons shows that the proposed framework can be extended to general machine learning problems in hyperbolic spaces.
arXiv Detail & Related papers (2022-03-07T21:36:21Z) - Highly Scalable and Provably Accurate Classification in Poincare Balls [40.82908295137667]
We establish a unified framework for learning scalable and simple hyperbolic linear classifiers with provable performance guarantees.
Our results include a new hyperbolic and second-order perceptron algorithm as well as an efficient and highly accurate convex optimization setup for hyperbolic support vector machine classifiers.
Their performance accuracies on synthetic data sets comprising millions of points, as well as on complex real-world data sets such as single-cell RNA-seq expression measurements, CIFAR10, Fashion-MNIST and mini-ImageNet.
arXiv Detail & Related papers (2021-09-08T16:59:39Z) - Augmented Sliced Wasserstein Distances [55.028065567756066]
We propose a new family of distance metrics, called augmented sliced Wasserstein distances (ASWDs)
ASWDs are constructed by first mapping samples to higher-dimensional hypersurfaces parameterized by neural networks.
Numerical results demonstrate that the ASWD significantly outperforms other Wasserstein variants for both synthetic and real-world problems.
arXiv Detail & Related papers (2020-06-15T23:00:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.