On the estimation of the number of components in multivariate functional principal component analysis
- URL: http://arxiv.org/abs/2311.04540v2
- Date: Fri, 12 Jul 2024 16:53:56 GMT
- Title: On the estimation of the number of components in multivariate functional principal component analysis
- Authors: Steven Golovkine, Edward Gunning, Andrew J. Simpkin, Norma Bargary,
- Abstract summary: We present extensive simulations to investigate choosing the number of principal components to retain.
We show empirically that the conventional approach of using a percentage of variance explained threshold for each univariate functional feature may be unreliable.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Happ and Greven (2018) developed a methodology for principal components analysis of multivariate functional data for data observed on different dimensional domains. Their approach relies on an estimation of univariate functional principal components for each univariate functional feature. In this paper, we present extensive simulations to investigate choosing the number of principal components to retain. We show empirically that the conventional approach of using a percentage of variance explained threshold for each univariate functional feature may be unreliable when aiming to explain an overall percentage of variance in the multivariate functional data, and thus we advise practitioners to be careful when using it.
Related papers
- Interpetable Target-Feature Aggregation for Multi-Task Learning based on Bias-Variance Analysis [53.38518232934096]
Multi-task learning (MTL) is a powerful machine learning paradigm designed to leverage shared knowledge across tasks to improve generalization and performance.
We propose an MTL approach at the intersection between task clustering and feature transformation based on a two-phase iterative aggregation of targets and features.
In both phases, a key aspect is to preserve the interpretability of the reduced targets and features through the aggregation with the mean, which is motivated by applications to Earth science.
arXiv Detail & Related papers (2024-06-12T08:30:16Z) - Regularized Multivariate Functional Principal Component Analysis [3.4238565157486187]
This paper introduces a novel approach called regularized theCA (ReCA) to address the issue of controlling the roughness of Principal Components.
The proposed method generates multivariate functional PCs, providing a concise and interpretable representation of the data.
arXiv Detail & Related papers (2023-06-24T14:22:25Z) - On the use of the Gram matrix for multivariate functional principal components analysis [0.0]
Dimension reduction is crucial in functional data analysis (FDA)
Existing approaches for functional principal component analysis usually involve the diagonalization of the covariance operator.
We propose to use the inner-product between the curves to estimate the eigenelements of multivariate and multidimensional functional datasets.
arXiv Detail & Related papers (2023-06-22T15:09:41Z) - Multi-task Bias-Variance Trade-off Through Functional Constraints [102.64082402388192]
Multi-task learning aims to acquire a set of functions that perform well for diverse tasks.
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks.
We introduce a constrained learning formulation that enforces domain specific solutions to a central function.
arXiv Detail & Related papers (2022-10-27T16:06:47Z) - Multivariate Wasserstein Functional Connectivity for Autism Screening [82.68524566142271]
We propose to compare regions of interest directly, without the use of representative time series.
We assess the proposed Wasserstein functional connectivity measure on the autism screening task.
arXiv Detail & Related papers (2022-09-23T16:23:05Z) - Embedding Functional Data: Multidimensional Scaling and Manifold
Learning [6.726255259929498]
We focus on classical scaling and Isomap -- prototypical methods that have played important roles in these area.
In the process, we highlight the crucial role that the ambient metric plays.
arXiv Detail & Related papers (2022-08-30T21:12:31Z) - Learning Functions on Multiple Sets using Multi-Set Transformers [31.09791656949115]
We show how to generalize this architecture to sets of elements of any dimension by dimension equivariance.
We demonstrate that our architecture is a universal approximator of these functions, and show superior results to existing methods on a variety of tasks.
arXiv Detail & Related papers (2022-06-30T17:39:15Z) - A geometric perspective on functional outlier detection [0.0]
We develop a conceptualization of functional outlier detection that is more widely applicable and realistic than previously proposed.
We show that simple manifold learning methods can be used to reliably infer and visualize the geometric structure of functional data sets.
Our experiments on synthetic and real data sets demonstrate that this approach leads to outlier detection performances at least on par with existing functional data-specific methods.
arXiv Detail & Related papers (2021-09-14T17:42:57Z) - Eigen Analysis of Self-Attention and its Reconstruction from Partial
Computation [58.80806716024701]
We study the global structure of attention scores computed using dot-product based self-attention.
We find that most of the variation among attention scores lie in a low-dimensional eigenspace.
We propose to compute scores only for a partial subset of token pairs, and use them to estimate scores for the remaining pairs.
arXiv Detail & Related papers (2021-06-16T14:38:42Z) - The role of feature space in atomistic learning [62.997667081978825]
Physically-inspired descriptors play a key role in the application of machine-learning techniques to atomistic simulations.
We introduce a framework to compare different sets of descriptors, and different ways of transforming them by means of metrics and kernels.
We compare representations built in terms of n-body correlations of the atom density, quantitatively assessing the information loss associated with the use of low-order features.
arXiv Detail & Related papers (2020-09-06T14:12:09Z) - Invariant Feature Coding using Tensor Product Representation [75.62232699377877]
We prove that the group-invariant feature vector contains sufficient discriminative information when learning a linear classifier.
A novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering.
arXiv Detail & Related papers (2019-06-05T07:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.