Learning Tree Structures from Leaves For Particle Decay Reconstruction
- URL: http://arxiv.org/abs/2208.14924v2
- Date: Thu, 1 Sep 2022 12:21:32 GMT
- Title: Learning Tree Structures from Leaves For Particle Decay Reconstruction
- Authors: James Kahn, Ilias Tsaklidis, Oskar Taubert, Lea Reuter, Giulio Dujany,
Tobias Boeckh, Arthur Thaller, Pablo Goldenzweig, Florian Bernlochner, Achim
Streit and Markus G\"otz
- Abstract summary: We present a neural approach to reconstructing rooted tree graphs describing hierarchical interactions, using a novel representation we term the Lowest Common Ancestor Generations (LCAG) matrix.
We are able to correctly predict the LCAG purely from leaf features for a maximum tree-depth of $8$ in $92.5%$ of cases for trees up to $6$ leaves (including) and $59.7%$ for trees up to $10$ in our simulated dataset.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we present a neural approach to reconstructing rooted tree
graphs describing hierarchical interactions, using a novel representation we
term the Lowest Common Ancestor Generations (LCAG) matrix. This compact
formulation is equivalent to the adjacency matrix, but enables learning a
tree's structure from its leaves alone without the prior assumptions required
if using the adjacency matrix directly. Employing the LCAG therefore enables
the first end-to-end trainable solution which learns the hierarchical structure
of varying tree sizes directly, using only the terminal tree leaves to do so.
In the case of high-energy particle physics, a particle decay forms a
hierarchical tree structure of which only the final products can be observed
experimentally, and the large combinatorial space of possible trees makes an
analytic solution intractable. We demonstrate the use of the LCAG as a target
in the task of predicting simulated particle physics decay structures using
both a Transformer encoder and a Neural Relational Inference encoder Graph
Neural Network. With this approach, we are able to correctly predict the LCAG
purely from leaf features for a maximum tree-depth of $8$ in $92.5\%$ of cases
for trees up to $6$ leaves (including) and $59.7\%$ for trees up to $10$ in our
simulated dataset.
Related papers
- Phylo2Vec: a vector representation for binary trees [0.49478969093606673]
We present Phylo2Vec, a parsimonious encoding for phylogenetic trees.
It serves as a unified approach for both manipulating and representing phylogenetic trees.
As a proof of concept, we use Phylo2Vec for maximum likelihood inference on five real-world datasets.
arXiv Detail & Related papers (2023-04-25T09:54:35Z) - Individualized and Global Feature Attributions for Gradient Boosted
Trees in the Presence of $\ell_2$ Regularization [0.0]
We propose Prediction Decomposition (PreDecomp), a novel individualized feature attribution for boosted trees when they are trained with $ell$ regularization.
We also propose TreeInner, a family of debiased global feature attributions defined in terms of the inner product between any individualized feature attribution and labels on out-sample data for each tree.
arXiv Detail & Related papers (2022-11-08T17:56:22Z) - A Neural Tangent Kernel Perspective of Infinite Tree Ensembles [8.020742121274417]
We introduce and study the Tree Neural Tangent Kernel (TNTK), which provides new insights into the behavior of the infinite ensemble of soft trees.
We find several non-trivial properties, such as the effect of the oblivious tree structure and the degeneracy of the TNTK induced by the deepening of the trees.
arXiv Detail & Related papers (2021-09-10T16:48:24Z) - Structural Optimization Makes Graph Classification Simpler and Better [5.770986723520119]
We investigate the feasibility of improving graph classification performance while simplifying the model learning process.
Inspired by progress in structural information assessment, we optimize the given data sample from graphs to encoding trees.
We present an implementation of the scheme in a tree kernel and a convolutional network to perform graph classification.
arXiv Detail & Related papers (2021-09-05T08:54:38Z) - Robustifying Algorithms of Learning Latent Trees with Vector Variables [92.18777020401484]
We present the sample complexities of Recursive Grouping (RG) and Chow-Liu Recursive Grouping (CLRG)
We robustify RG, CLRG, Neighbor Joining (NJ) and Spectral NJ (SNJ) by using the truncated inner product.
We derive the first known instance-dependent impossibility result for structure learning of latent trees.
arXiv Detail & Related papers (2021-06-02T01:37:52Z) - Spectral Top-Down Recovery of Latent Tree Models [13.681975313065477]
Spectral Top-Down Recovery (STDR) is a divide-and-conquer approach for inference of large latent tree models.
STDR's partitioning step is non-random. Instead, it is based on the Fiedler vector of a suitable Laplacian matrix related to the observed nodes.
We prove that STDR is statistically consistent, and bound the number of samples required to accurately recover the tree with high probability.
arXiv Detail & Related papers (2021-02-26T02:47:42Z) - Visualizing hierarchies in scRNA-seq data using a density tree-biased
autoencoder [50.591267188664666]
We propose an approach for identifying a meaningful tree structure from high-dimensional scRNA-seq data.
We then introduce DTAE, a tree-biased autoencoder that emphasizes the tree structure of the data in low dimensional space.
arXiv Detail & Related papers (2021-02-11T08:48:48Z) - SGA: A Robust Algorithm for Partial Recovery of Tree-Structured
Graphical Models with Noisy Samples [75.32013242448151]
We consider learning Ising tree models when the observations from the nodes are corrupted by independent but non-identically distributed noise.
Katiyar et al. (2020) showed that although the exact tree structure cannot be recovered, one can recover a partial tree structure.
We propose Symmetrized Geometric Averaging (SGA), a more statistically robust algorithm for partial tree recovery.
arXiv Detail & Related papers (2021-01-22T01:57:35Z) - Growing Deep Forests Efficiently with Soft Routing and Learned
Connectivity [79.83903179393164]
This paper further extends the deep forest idea in several important aspects.
We employ a probabilistic tree whose nodes make probabilistic routing decisions, a.k.a., soft routing, rather than hard binary decisions.
Experiments on the MNIST dataset demonstrate that our empowered deep forests can achieve better or comparable performance than [1],[3].
arXiv Detail & Related papers (2020-12-29T18:05:05Z) - Rethinking Learnable Tree Filter for Generic Feature Transform [71.77463476808585]
Learnable Tree Filter presents a remarkable approach to model structure-preserving relations for semantic segmentation.
To relax the geometric constraint, we give the analysis by reformulating it as a Markov Random Field and introduce a learnable unary term.
For semantic segmentation, we achieve leading performance (82.1% mIoU) on the Cityscapes benchmark without bells-and-whistles.
arXiv Detail & Related papers (2020-12-07T07:16:47Z) - Recursive Top-Down Production for Sentence Generation with Latent Trees [77.56794870399288]
We model the production property of context-free grammars for natural and synthetic languages.
We present a dynamic programming algorithm that marginalises over latent binary tree structures with $N$ leaves.
We also present experimental results on German-English translation on the Multi30k dataset.
arXiv Detail & Related papers (2020-10-09T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.