Dimension Reduction and Data Visualization for Fr\'echet Regression
- URL: http://arxiv.org/abs/2110.00467v1
- Date: Fri, 1 Oct 2021 15:01:32 GMT
- Title: Dimension Reduction and Data Visualization for Fr\'echet Regression
- Authors: Qi Zhang, Lingzhou Xue, and Bing Li
- Abstract summary: Fr'echet regression model provides a promising framework for regression analysis with metric spacevalued responses.
We introduce a flexible sufficient dimension reduction (SDR) method for Fr'echet regression to achieve two purposes.
- Score: 8.713190936209156
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rapid development of data collection techniques, complex data
objects that are not in the Euclidean space are frequently encountered in new
statistical applications. Fr\'echet regression model (Peterson & M\"uller 2019)
provides a promising framework for regression analysis with metric space-valued
responses. In this paper, we introduce a flexible sufficient dimension
reduction (SDR) method for Fr\'echet regression to achieve two purposes: to
mitigate the curse of dimensionality caused by high-dimensional predictors, and
to provide a tool for data visualization for Fr\'echet regression. Our approach
is flexible enough to turn any existing SDR method for Euclidean (X,Y) into one
for Euclidean X and metric space-valued Y. The basic idea is to first map the
metric-space valued random object $Y$ to a real-valued random variable $f(Y)$
using a class of functions, and then perform classical SDR to the transformed
data. If the class of functions is sufficiently rich, then we are guaranteed to
uncover the Fr\'echet SDR space. We showed that such a class, which we call an
ensemble, can be generated by a universal kernel. We established the
consistency and asymptotic convergence rate of the proposed methods. The
finite-sample performance of the proposed methods is illustrated through
simulation studies for several commonly encountered metric spaces that include
Wasserstein space, the space of symmetric positive definite matrices, and the
sphere. We illustrated the data visualization aspect of our method by exploring
the human mortality distribution data across countries and by studying the
distribution of hematoma density.
Related papers
- Deep Fréchet Regression [4.915744683251151]
We propose a flexible regression model capable of handling high-dimensional predictors without imposing parametric assumptions.
The proposed model outperforms existing methods for non-Euclidean responses.
arXiv Detail & Related papers (2024-07-31T07:54:14Z) - An Optimal Transport Approach for Network Regression [0.6238182916866519]
We build upon recent developments in generalized regression models on metric spaces based on Fr'echet means.
We propose a network regression method using the Wasserstein metric.
arXiv Detail & Related papers (2024-06-18T02:03:07Z) - Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Federated Sufficient Dimension Reduction Through High-Dimensional Sparse
Sliced Inverse Regression [4.561305216067566]
Federated learning has become a popular tool in the big data era nowadays.
We propose a federated sparse sliced inverse regression algorithm for the first time.
arXiv Detail & Related papers (2023-01-23T15:53:06Z) - Bayesian Hyperbolic Multidimensional Scaling [2.5944208050492183]
We propose a Bayesian approach to multidimensional scaling when the low-dimensional manifold is hyperbolic.
A case-control likelihood approximation allows for efficient sampling from the posterior distribution in larger data settings.
We evaluate the proposed method against state-of-the-art alternatives using simulations, canonical reference datasets, Indian village network data, and human gene expression data.
arXiv Detail & Related papers (2022-10-26T23:34:30Z) - Distributed Dynamic Safe Screening Algorithms for Sparse Regularization [73.85961005970222]
We propose a new distributed dynamic safe screening (DDSS) method for sparsity regularized models and apply it on shared-memory and distributed-memory architecture respectively.
We prove that the proposed method achieves the linear convergence rate with lower overall complexity and can eliminate almost all the inactive features in a finite number of iterations almost surely.
arXiv Detail & Related papers (2022-04-23T02:45:55Z) - Random Forest Weighted Local Fréchet Regression with Random Objects [18.128663071848923]
We propose a novel random forest weighted local Fr'echet regression paradigm.
Our first method uses these weights as the local average to solve the conditional Fr'echet mean.
Second method performs local linear Fr'echet regression, both significantly improving existing Fr'echet regression methods.
arXiv Detail & Related papers (2022-02-10T09:10:59Z) - Manifold Hypothesis in Data Analysis: Double Geometrically-Probabilistic
Approach to Manifold Dimension Estimation [92.81218653234669]
We present new approach to manifold hypothesis checking and underlying manifold dimension estimation.
Our geometrical method is a modification for sparse data of a well-known box-counting algorithm for Minkowski dimension calculation.
Experiments on real datasets show that the suggested approach based on two methods combination is powerful and effective.
arXiv Detail & Related papers (2021-07-08T15:35:54Z) - Deep Dimension Reduction for Supervised Representation Learning [51.10448064423656]
We propose a deep dimension reduction approach to learning representations with essential characteristics.
The proposed approach is a nonparametric generalization of the sufficient dimension reduction method.
We show that the estimated deep nonparametric representation is consistent in the sense that its excess risk converges to zero.
arXiv Detail & Related papers (2020-06-10T14:47:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.