Machine Learning Assisted Orthonormal Basis Selection for Functional
Data Analysis
- URL: http://arxiv.org/abs/2103.07453v1
- Date: Fri, 12 Mar 2021 18:27:29 GMT
- Title: Machine Learning Assisted Orthonormal Basis Selection for Functional
Data Analysis
- Authors: Rani Basna, Hiba Nassar and Krzysztof Podg\'orski
- Abstract summary: We propose a strictly data-driven method of orthonormal basis selection.
The algorithm learns from the data in the machine learning style to efficiently place knots.
The optimality criterion is based on the average (per functional data point) mean square error and is utilized both in the learning algorithms and in comparison studies.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In implementations of the functional data methods, the effect of the initial
choice of an orthonormal basis has not gained much attention in the past.
Typically, several standard bases such as Fourier, wavelets, splines, etc. are
considered to transform observed functional data and a choice is made without
any formal criteria indicating which of the bases is preferable for the initial
transformation of the data into functions. In an attempt to address this issue,
we propose a strictly data-driven method of orthogonal basis selection. The
method uses recently introduced orthogonal spline bases called the splinets
obtained by efficient orthogonalization of the B-splines. The algorithm learns
from the data in the machine learning style to efficiently place knots. The
optimality criterion is based on the average (per functional data point) mean
square error and is utilized both in the learning algorithms and in comparison
studies. The latter indicates efficiency that is particularly evident for the
sparse functional data and to a lesser degree in analyses of responses to
complex physical systems.
Related papers
- DiffuMatch: Category-Agnostic Spectral Diffusion Priors for Robust Non-rigid Shape Matching [53.39693288324375]
We show that both in-network regularization and functional map training can be replaced with data-driven methods.<n>We first train a generative model of functional maps in the spectral domain using score-based generative modeling.<n>We then exploit the resulting model to promote the structural properties of ground truth functional maps on new shape collections.
arXiv Detail & Related papers (2025-07-31T16:44:54Z) - Discovering Interpretable Ordinary Differential Equations from Noisy Data [0.0]
We propose an unsupervised parameter estimation methodology that finds an approximate general solution, followed by a spline transformation to linearly estimate the coefficients of the governing ordinary differential equation (ODE)<n>The approximate general solution is postulated using the same functional form as the analytical solution of a general homogeneous, linear, constant-coefficient ODE.<n>The methodology is also robust to noisy data and thus allows the integration of data-driven techniques into real experimental setting for data-driven learning of physical phenomena.
arXiv Detail & Related papers (2025-07-29T14:19:59Z) - Generalized Tensor-based Parameter-Efficient Fine-Tuning via Lie Group Transformations [50.010924231754856]
Adapting pre-trained foundation models for diverse downstream tasks is a core practice in artificial intelligence.
To overcome this, parameter-efficient fine-tuning (PEFT) methods like LoRA have emerged and are becoming a growing research focus.
We propose a generalization that extends matrix-based PEFT methods to higher-dimensional parameter spaces without compromising their structural properties.
arXiv Detail & Related papers (2025-04-01T14:36:45Z) - Learning dynamical systems from data: Gradient-based dictionary optimization [0.8643517734716606]
We present a novel gradient descent-based optimization framework for learning suitable basis functions from data.
We show how it can be used in combination with EDMD, SINDy, and PDE-FIND.
arXiv Detail & Related papers (2024-11-07T15:15:27Z) - Symmetry Discovery for Different Data Types [52.2614860099811]
Equivariant neural networks incorporate symmetries into their architecture, achieving higher generalization performance.
We propose LieSD, a method for discovering symmetries via trained neural networks which approximate the input-output mappings of the tasks.
We validate the performance of LieSD on tasks with symmetries such as the two-body problem, the moment of inertia matrix prediction, and top quark tagging.
arXiv Detail & Related papers (2024-10-13T13:39:39Z) - Functional Autoencoder for Smoothing and Representation Learning [0.0]
We propose to learn the nonlinear representations of functional data using neural network autoencoders designed to process data in the form it is usually collected without the need of preprocessing.
We design the encoder to employ a projection layer computing the weighted inner product of the functional data and functional weights over the observed timestamp, and the decoder to apply a recovery layer that maps the finite-dimensional vector extracted from the functional data back to functional space.
arXiv Detail & Related papers (2024-01-17T08:33:25Z) - Gram-Schmidt Methods for Unsupervised Feature Extraction and Selection [7.373617024876725]
We propose a Gram-Schmidt process over function spaces to detect and map out nonlinear dependencies.
We provide experimental results for synthetic and real-world benchmark datasets.
Surprisingly, our linear feature extraction algorithms are comparable and often outperform several important nonlinear feature extraction methods.
arXiv Detail & Related papers (2023-11-15T21:29:57Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Efficient Multidimensional Functional Data Analysis Using Marginal
Product Basis Systems [2.4554686192257424]
We propose a framework for learning continuous representations from a sample of multidimensional functional data.
We show that the resulting estimation problem can be solved efficiently by the tensor decomposition.
We conclude with a real data application in neuroimaging.
arXiv Detail & Related papers (2021-07-30T16:02:15Z) - Functional Classwise Principal Component Analysis: A Novel
Classification Framework [4.6592517049808455]
We present a novel classification framework using functional data and classwise Principal Component Analysis (PCA)
Our method extracts a piece wise linear functional feature space and is particularly suitable for hard classification problems.
We demonstrate the efficacy of our proposed method by applying it to both synthetic data sets and real time series data from diverse fields.
arXiv Detail & Related papers (2021-06-26T07:10:58Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - Improved guarantees and a multiple-descent curve for Column Subset
Selection and the Nystr\"om method [76.73096213472897]
We develop techniques which exploit spectral properties of the data matrix to obtain improved approximation guarantees.
Our approach leads to significantly better bounds for datasets with known rates of singular value decay.
We show that both our improved bounds and the multiple-descent curve can be observed on real datasets simply by varying the RBF parameter.
arXiv Detail & Related papers (2020-02-21T00:43:06Z) - Supervised Quantile Normalization for Low-rank Matrix Approximation [50.445371939523305]
We learn the parameters of quantile normalization operators that can operate row-wise on the values of $X$ and/or of its factorization $UV$ to improve the quality of the low-rank representation of $X$ itself.
We demonstrate the applicability of these techniques on synthetic and genomics datasets.
arXiv Detail & Related papers (2020-02-08T21:06:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.