Leonardo vindicated: Pythagorean trees for minimal reconstruction of the natural branching structures
- URL: http://arxiv.org/abs/2411.08024v1
- Date: Tue, 12 Nov 2024 18:54:55 GMT
- Title: Leonardo vindicated: Pythagorean trees for minimal reconstruction of the natural branching structures
- Authors: Dymitr Ruta, Corrado Mio, Ernesto Damiani,
- Abstract summary: Pythagorean tree is a fractal design that realistically mimics the natural tree branching structures.
We study various types of Pythagorean-like fractal trees with different shapes of the base, branching angles and relaxed scales.
We have developed a flexibly parameterised and fast algorithm to grow and visually examine deep Pythagorean-inspired fractal trees.
- Score: 2.620755151566635
- License:
- Abstract: Trees continue to fascinate with their natural beauty and as engineering masterpieces optimal with respect to several independent criteria. Pythagorean tree is a well-known fractal design that realistically mimics the natural tree branching structures. We study various types of Pythagorean-like fractal trees with different shapes of the base, branching angles and relaxed scales in an attempt to identify and explain which variants are the closest match to the branching structures commonly observed in the natural world. Pursuing simultaneously the realism and minimalism of the fractal tree model, we have developed a flexibly parameterised and fast algorithm to grow and visually examine deep Pythagorean-inspired fractal trees with the capability to orderly over- or underestimate the Leonardo da Vinci's tree branching rule as well as control various imbalances and branching angles. We tested the realism of the generated fractal tree images by means of the classification accuracy of detecting natural tree with the transfer-trained deep Convolutional Neural Networks (CNNs). Having empirically established the parameters of the fractal trees that maximize the CNN's natural tree class classification accuracy we have translated them back to the scales and angles of branches and came to the interesting conclusions that support the da Vinci branching rule and golden ratio based scaling for both the shape of the branch and imbalance between the child branches, and claim the flexibly parameterized fractal trees can be used to generate artificial examples to train robust detectors of different species of trees.
Related papers
- $TrIND$: Representing Anatomical Trees by Denoising Diffusion of Implicit Neural Fields [17.943355593568242]
Anatomical trees play a central role in clinical diagnosis and treatment planning.
Traditional methods for representing tree structures exhibit drawbacks in terms of resolution, flexibility, and efficiency.
We propose a novel approach, $TrIND$, for representing anatomical trees using implicit neural representations.
arXiv Detail & Related papers (2024-03-13T21:43:24Z) - Learning a Decision Tree Algorithm with Transformers [75.96920867382859]
We introduce MetaTree, a transformer-based model trained via meta-learning to directly produce strong decision trees.
We fit both greedy decision trees and globally optimized decision trees on a large number of datasets, and train MetaTree to produce only the trees that achieve strong generalization performance.
arXiv Detail & Related papers (2024-02-06T07:40:53Z) - Why do Random Forests Work? Understanding Tree Ensembles as
Self-Regularizing Adaptive Smoothers [68.76846801719095]
We argue that the current high-level dichotomy into bias- and variance-reduction prevalent in statistics is insufficient to understand tree ensembles.
We show that forests can improve upon trees by three distinct mechanisms that are usually implicitly entangled.
arXiv Detail & Related papers (2024-02-02T15:36:43Z) - PhyloGFN: Phylogenetic inference with generative flow networks [57.104166650526416]
We introduce the framework of generative flow networks (GFlowNets) to tackle two core problems in phylogenetics: parsimony-based and phylogenetic inference.
Because GFlowNets are well-suited for sampling complex structures, they are a natural choice for exploring and sampling from the multimodal posterior distribution over tree topologies.
We demonstrate that our amortized posterior sampler, PhyloGFN, produces diverse and high-quality evolutionary hypotheses on real benchmark datasets.
arXiv Detail & Related papers (2023-10-12T23:46:08Z) - DeepTree: Modeling Trees with Situated Latents [8.372189962601073]
We propose a novel method for modeling trees based on learning developmental rules for branching structures instead of manually defining them.
We call our deep neural model situated latent because its behavior is determined by the intrinsic state.
Our method enables generating a wide variety of tree shapes without the need to define intricate parameters.
arXiv Detail & Related papers (2023-05-09T03:33:14Z) - Characterizing Intrinsic Compositionality in Transformers with Tree
Projections [72.45375959893218]
neural models like transformers can route information arbitrarily between different parts of their input.
We show that transformers for three different tasks become more treelike over the course of training.
These trees are predictive of model behavior, with more tree-like models generalizing better on tests of compositional generalization.
arXiv Detail & Related papers (2022-11-02T17:10:07Z) - Some theoretical results on discrete contour trees [3.1963310035729657]
We define a discrete contour tree, called the iso-tree, on a scalar graph.
We show that the iso-tree model works for data of all dimensions, and develop an axiomatic system formalizing the discrete contour structures.
arXiv Detail & Related papers (2022-06-24T07:31:11Z) - Tree Reconstruction using Topology Optimisation [0.685316573653194]
We present a general method for extracting the branch structure of trees from point cloud data.
We discuss the benefits and drawbacks of this novel approach to tree structure reconstruction.
Our method generates detailed and accurate tree structures in most cases.
arXiv Detail & Related papers (2022-05-26T07:08:32Z) - A Neural Tangent Kernel Perspective of Infinite Tree Ensembles [8.020742121274417]
We introduce and study the Tree Neural Tangent Kernel (TNTK), which provides new insights into the behavior of the infinite ensemble of soft trees.
We find several non-trivial properties, such as the effect of the oblivious tree structure and the degeneracy of the TNTK induced by the deepening of the trees.
arXiv Detail & Related papers (2021-09-10T16:48:24Z) - SGA: A Robust Algorithm for Partial Recovery of Tree-Structured
Graphical Models with Noisy Samples [75.32013242448151]
We consider learning Ising tree models when the observations from the nodes are corrupted by independent but non-identically distributed noise.
Katiyar et al. (2020) showed that although the exact tree structure cannot be recovered, one can recover a partial tree structure.
We propose Symmetrized Geometric Averaging (SGA), a more statistically robust algorithm for partial tree recovery.
arXiv Detail & Related papers (2021-01-22T01:57:35Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.