A Bayesian Framework on Asymmetric Mixture of Factor Analyser
- URL: http://arxiv.org/abs/2211.00729v1
- Date: Tue, 1 Nov 2022 20:19:52 GMT
- Title: A Bayesian Framework on Asymmetric Mixture of Factor Analyser
- Authors: Hamid Reza Safaeyan, Karim Zare, Mohamad R. Mahmoudi, Amir Mosavi
- Abstract summary: This paper introduces an MFA model with a rich and flexible class of skew normal (unrestricted) generalized hyperbolic (called SUNGH) distributions.
The SUNGH family provides considerable flexibility to model skewness in different directions as well as allowing for heavy tailed data.
Considering factor analysis models, the SUNGH family also allows for skewness and heavy tails for both the error component and factor scores.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mixture of factor analyzer (MFA) model is an efficient model for the analysis
of high dimensional data through which the factor-analyzer technique based on
the covariance matrices reducing the number of free parameters. The model also
provides an important methodology to determine latent groups in data. There are
several pieces of research to extend the model based on the asymmetrical and/or
with outlier datasets with some known computational limitations that have been
examined in frequentist cases. In this paper, an MFA model with a rich and
flexible class of skew normal (unrestricted) generalized hyperbolic (called
SUNGH) distributions along with a Bayesian structure with several computational
benefits have been introduced. The SUNGH family provides considerable
flexibility to model skewness in different directions as well as allowing for
heavy tailed data. There are several desirable properties in the structure of
the SUNGH family, including, an analytically flexible density which leads to
easing up the computation applied for the estimation of parameters. Considering
factor analysis models, the SUNGH family also allows for skewness and heavy
tails for both the error component and factor scores. In the present study, the
advantages of using this family of distributions have been discussed and the
suitable efficiency of the introduced MFA model using real data examples and
simulation has been demonstrated.
Related papers
- Factor Adjusted Spectral Clustering for Mixture Models [6.806940901668607]
We propose the Factor Adjusted Spectral Clustering (FASC) algorithm, which uses an additional data denoising step via eliminating the factor component to cope with the data dependency.
Our assumption bridges many classical factor models in the literature, such as the pervasive factor model, the weak factor model, and the sparse factor model.
FASC provides significant results in many cases where traditional spectral clustering fails.
arXiv Detail & Related papers (2024-08-22T17:31:21Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Finite Mixtures of Multivariate Poisson-Log Normal Factor Analyzers for
Clustering Count Data [0.8499685241219366]
A class of eight parsimonious mixture models based on the mixtures of factor analyzers model are introduced.
The proposed models are explored in the context of clustering discrete data arising from RNA sequencing studies.
arXiv Detail & Related papers (2023-11-13T21:23:15Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Nonparametric Functional Analysis of Generalized Linear Models Under
Nonlinear Constraints [0.0]
This article introduces a novel nonparametric methodology for Generalized Linear Models.
It combines the strengths of the binary regression and latent variable formulations for categorical data.
It extends recently published parametric versions of the methodology and generalizes it.
arXiv Detail & Related papers (2021-10-11T04:49:59Z) - Generative Learning of Heterogeneous Tail Dependence [13.60514494665717]
Our model features heterogeneous and asymmetric tail dependence between all pairs of individual dimensions.
We devise a novel moment learning algorithm to learn the parameters.
Results show that this framework gives better finite-sample performance compared to the copula-based benchmarks.
arXiv Detail & Related papers (2020-11-26T05:34:31Z) - Generalized Matrix Factorization: efficient algorithms for fitting
generalized linear latent variable models to large data arrays [62.997667081978825]
Generalized Linear Latent Variable models (GLLVMs) generalize such factor models to non-Gaussian responses.
Current algorithms for estimating model parameters in GLLVMs require intensive computation and do not scale to large datasets.
We propose a new approach for fitting GLLVMs to high-dimensional datasets, based on approximating the model using penalized quasi-likelihood.
arXiv Detail & Related papers (2020-10-06T04:28:19Z) - Bayesian Sparse Factor Analysis with Kernelized Observations [67.60224656603823]
Multi-view problems can be faced with latent variable models.
High-dimensionality and non-linear issues are traditionally handled by kernel methods.
We propose merging both approaches into single model.
arXiv Detail & Related papers (2020-06-01T14:25:38Z) - Predicting Multidimensional Data via Tensor Learning [0.0]
We develop a model that retains the intrinsic multidimensional structure of the dataset.
To estimate the model parameters, an Alternating Least Squares algorithm is developed.
The proposed model is able to outperform benchmark models present in the forecasting literature.
arXiv Detail & Related papers (2020-02-11T11:57:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.