Optimal estimation of Gaussian (poly)trees
- URL: http://arxiv.org/abs/2402.06380v1
- Date: Fri, 9 Feb 2024 12:58:36 GMT
- Title: Optimal estimation of Gaussian (poly)trees
- Authors: Yuhao Wang, Ming Gao, Wai Ming Tai, Bryon Aragam, Arnab Bhattacharyya
- Abstract summary: We consider both problems of distribution learning (i.e. in KL distance) and structure learning (i.e. exact recovery)
The first approach is based on the Chow-Liu algorithm, and learns an optimal tree-structured distribution efficiently.
The second approach is a modification of the PC algorithm for polytrees that uses partial correlation as a conditional independence tester for constraint-based structure learning.
- Score: 25.02920605955238
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We develop optimal algorithms for learning undirected Gaussian trees and
directed Gaussian polytrees from data. We consider both problems of
distribution learning (i.e. in KL distance) and structure learning (i.e. exact
recovery). The first approach is based on the Chow-Liu algorithm, and learns an
optimal tree-structured distribution efficiently. The second approach is a
modification of the PC algorithm for polytrees that uses partial correlation as
a conditional independence tester for constraint-based structure learning. We
derive explicit finite-sample guarantees for both approaches, and show that
both approaches are optimal by deriving matching lower bounds. Additionally, we
conduct numerical experiments to compare the performance of various algorithms,
providing further insights and empirical evidence.
Related papers
- Learning Linear Gaussian Polytree Models with Interventions [0.0]
We present a consistent and highly scalable local approach to learn the causal structure of a linear Gaussian polytree.
Our methods first learn the skeleton of the polytree and then orient its edges.
Our simulation studies demonstrate that our approach is fast, has good accuracy in terms of structural Hamming distance, and handles problems with thousands of nodes.
arXiv Detail & Related papers (2023-11-08T12:29:19Z) - Rolling Lookahead Learning for Optimal Classification Trees [0.0]
We propose a rolling subtree lookahead algorithm that combines the relative scalability of the myopic approaches with the foresight of the optimal approaches in constructing trees.
We show that our approach outperforms optimal and myopic approaches in 808 out of 1330 problem instances, improving the out-of-sample accuracy by up to 23.6% and 14.4%, respectively.
arXiv Detail & Related papers (2023-04-21T09:17:00Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - Multivariate Systemic Risk Measures and Computation by Deep Learning
Algorithms [63.03966552670014]
We discuss the key related theoretical aspects, with a particular focus on the fairness properties of primal optima and associated risk allocations.
The algorithms we provide allow for learning primals, optima for the dual representation and corresponding fair risk allocations.
arXiv Detail & Related papers (2023-02-02T22:16:49Z) - Learning Linear Non-Gaussian Polytree Models [2.4493299476776778]
We propose new algorithms to efficiently learn graphs that are polytrees.
Our approach combines the Chow--Liu algorithm, which first learns the undirected tree structure, with novel schemes to orient the edges.
arXiv Detail & Related papers (2022-08-13T18:20:10Z) - Tree ensemble kernels for Bayesian optimization with known constraints
over mixed-feature spaces [54.58348769621782]
Tree ensembles can be well-suited for black-box optimization tasks such as algorithm tuning and neural architecture search.
Two well-known challenges in using tree ensembles for black-box optimization are (i) effectively quantifying model uncertainty for exploration and (ii) optimizing over the piece-wise constant acquisition function.
Our framework performs as well as state-of-the-art methods for unconstrained black-box optimization over continuous/discrete features and outperforms competing methods for problems combining mixed-variable feature spaces and known input constraints.
arXiv Detail & Related papers (2022-07-02T16:59:37Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - Bilevel Optimization: Convergence Analysis and Enhanced Design [63.64636047748605]
Bilevel optimization is a tool for many machine learning problems.
We propose a novel stoc-efficientgradient estimator named stoc-BiO.
arXiv Detail & Related papers (2020-10-15T18:09:48Z) - Efficient Computation of Expectations under Spanning Tree Distributions [67.71280539312536]
We propose unified algorithms for the important cases of first-order expectations and second-order expectations in edge-factored, non-projective spanning-tree models.
Our algorithms exploit a fundamental connection between gradients and expectations, which allows us to derive efficient algorithms.
arXiv Detail & Related papers (2020-08-29T14:58:26Z) - Evolutionary algorithms for constructing an ensemble of decision trees [0.0]
We propose several methods for induction of decision trees and their ensembles based on evolutionary algorithms.
The main difference of our approach is using real-valued vector representation of decision tree.
We test the predictive performance of this methods using several public UCI data sets.
arXiv Detail & Related papers (2020-02-03T13:38:50Z) - Boosting Algorithms for Estimating Optimal Individualized Treatment
Rules [4.898659895355356]
We present nonparametric algorithms for estimating optimal individualized treatment rules.
The proposed algorithms are based on the XGBoost algorithm, which is known as one of the most powerful algorithms in the machine learning literature.
arXiv Detail & Related papers (2020-01-31T22:26:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.