On the estimation of the number of components in multivariate functional principal component analysis
- URL: http://arxiv.org/abs/2311.04540v2
- Date: Fri, 12 Jul 2024 16:53:56 GMT
- Title: On the estimation of the number of components in multivariate functional principal component analysis
- Authors: Steven Golovkine, Edward Gunning, Andrew J. Simpkin, Norma Bargary,
- Abstract summary: We present extensive simulations to investigate choosing the number of principal components to retain.
We show empirically that the conventional approach of using a percentage of variance explained threshold for each univariate functional feature may be unreliable.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Happ and Greven (2018) developed a methodology for principal components analysis of multivariate functional data for data observed on different dimensional domains. Their approach relies on an estimation of univariate functional principal components for each univariate functional feature. In this paper, we present extensive simulations to investigate choosing the number of principal components to retain. We show empirically that the conventional approach of using a percentage of variance explained threshold for each univariate functional feature may be unreliable when aiming to explain an overall percentage of variance in the multivariate functional data, and thus we advise practitioners to be careful when using it.
Related papers
- Disentangling Interactions and Dependencies in Feature Attribution [9.442326245744916]
In machine learning, global feature importance methods try to determine how much each individual feature contributes to predicting a target variable.
In commonly used feature importance scores these cooperative effects are conflated with the features' individual contributions.
We derive DIP, a new mathematical decomposition of individual feature importance scores that disentangles three components.
arXiv Detail & Related papers (2024-10-31T09:41:10Z) - Augmented Functional Random Forests: Classifier Construction and Unbiased Functional Principal Components Importance through Ad-Hoc Conditional Permutations [0.0]
This paper introduces a novel supervised classification strategy that integrates functional data analysis with tree-based methods.
We propose augmented versions of functional classification trees and functional random forests, incorporating a new tool for assessing the importance of functional principal components.
arXiv Detail & Related papers (2024-08-23T15:58:41Z) - Partial-Multivariate Model for Forecasting [28.120094495344453]
We propose a Transformer-based partial-multivariate model, PMformer, for forecasting problems.
We demonstrate that PMformer outperforms various univariate and complete-multivariate models.
We also highlight other advantages of PMformer: efficiency and robustness under missing features.
arXiv Detail & Related papers (2024-08-19T05:18:50Z) - Interpetable Target-Feature Aggregation for Multi-Task Learning based on Bias-Variance Analysis [53.38518232934096]
Multi-task learning (MTL) is a powerful machine learning paradigm designed to leverage shared knowledge across tasks to improve generalization and performance.
We propose an MTL approach at the intersection between task clustering and feature transformation based on a two-phase iterative aggregation of targets and features.
In both phases, a key aspect is to preserve the interpretability of the reduced targets and features through the aggregation with the mean, which is motivated by applications to Earth science.
arXiv Detail & Related papers (2024-06-12T08:30:16Z) - Regularized Multivariate Functional Principal Component Analysis [3.4238565157486187]
This paper introduces a novel approach called regularized theCA (ReCA) to address the issue of controlling the roughness of Principal Components.
The proposed method generates multivariate functional PCs, providing a concise and interpretable representation of the data.
arXiv Detail & Related papers (2023-06-24T14:22:25Z) - On the use of the Gram matrix for multivariate functional principal components analysis [0.0]
Dimension reduction is crucial in functional data analysis (FDA)
Existing approaches for functional principal component analysis usually involve the diagonalization of the covariance operator.
We propose to use the inner-product between the curves to estimate the eigenelements of multivariate and multidimensional functional datasets.
arXiv Detail & Related papers (2023-06-22T15:09:41Z) - Multi-task Bias-Variance Trade-off Through Functional Constraints [102.64082402388192]
Multi-task learning aims to acquire a set of functions that perform well for diverse tasks.
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks.
We introduce a constrained learning formulation that enforces domain specific solutions to a central function.
arXiv Detail & Related papers (2022-10-27T16:06:47Z) - Multivariate Wasserstein Functional Connectivity for Autism Screening [82.68524566142271]
We propose to compare regions of interest directly, without the use of representative time series.
We assess the proposed Wasserstein functional connectivity measure on the autism screening task.
arXiv Detail & Related papers (2022-09-23T16:23:05Z) - Eigen Analysis of Self-Attention and its Reconstruction from Partial
Computation [58.80806716024701]
We study the global structure of attention scores computed using dot-product based self-attention.
We find that most of the variation among attention scores lie in a low-dimensional eigenspace.
We propose to compute scores only for a partial subset of token pairs, and use them to estimate scores for the remaining pairs.
arXiv Detail & Related papers (2021-06-16T14:38:42Z) - The role of feature space in atomistic learning [62.997667081978825]
Physically-inspired descriptors play a key role in the application of machine-learning techniques to atomistic simulations.
We introduce a framework to compare different sets of descriptors, and different ways of transforming them by means of metrics and kernels.
We compare representations built in terms of n-body correlations of the atom density, quantitatively assessing the information loss associated with the use of low-order features.
arXiv Detail & Related papers (2020-09-06T14:12:09Z) - Invariant Feature Coding using Tensor Product Representation [75.62232699377877]
We prove that the group-invariant feature vector contains sufficient discriminative information when learning a linear classifier.
A novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering.
arXiv Detail & Related papers (2019-06-05T07:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.