A new computationally efficient algorithm to solve Feature Selection for
Functional Data Classification in high-dimensional spaces
- URL: http://arxiv.org/abs/2401.05765v2
- Date: Tue, 5 Mar 2024 10:48:22 GMT
- Title: A new computationally efficient algorithm to solve Feature Selection for
Functional Data Classification in high-dimensional spaces
- Authors: Tobia Boschi, Francesca Bonin, Rodrigo Ordonez-Hurtado, Alessandra
Pascale, and Jonathan Epperlein
- Abstract summary: This paper introduces a novel methodology for Feature Selection for Functional Classification, FSFC, that addresses the challenge of jointly performing feature selection and classification of functional data.
We employ functional principal components and develop a new adaptive version of the Dual Augmented Lagrangian algorithm.
The computational efficiency of FSFC enables handling high-dimensional scenarios where the number of features may considerably exceed the number of statistical units.
- Score: 41.79972837288401
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces a novel methodology for Feature Selection for
Functional Classification, FSFC, that addresses the challenge of jointly
performing feature selection and classification of functional data in scenarios
with categorical responses and multivariate longitudinal features. FSFC tackles
a newly defined optimization problem that integrates logistic loss and
functional features to identify the most crucial variables for classification.
To address the minimization procedure, we employ functional principal
components and develop a new adaptive version of the Dual Augmented Lagrangian
algorithm. The computational efficiency of FSFC enables handling
high-dimensional scenarios where the number of features may considerably exceed
the number of statistical units. Simulation experiments demonstrate that FSFC
outperforms other machine learning and deep learning methods in computational
time and classification accuracy. Furthermore, the FSFC feature selection
capability can be leveraged to significantly reduce the problem's
dimensionality and enhance the performances of other classification algorithms.
The efficacy of FSFC is also demonstrated through a real data application,
analyzing relationships between four chronic diseases and other health and
demographic factors.
Related papers
- A Contrast Based Feature Selection Algorithm for High-dimensional Data
set in Machine Learning [9.596923373834093]
We propose a novel filter feature selection method, ContrastFS, which selects discriminative features based on the discrepancies features shown between different classes.
We validate effectiveness and efficiency of our approach on several widely studied benchmark datasets, results show that the new method performs favorably with negligible computation.
arXiv Detail & Related papers (2024-01-15T05:32:35Z) - Analysis and Optimization of Wireless Federated Learning with Data
Heterogeneity [72.85248553787538]
This paper focuses on performance analysis and optimization for wireless FL, considering data heterogeneity, combined with wireless resource allocation.
We formulate the loss function minimization problem, under constraints on long-term energy consumption and latency, and jointly optimize client scheduling, resource allocation, and the number of local training epochs (CRE)
Experiments on real-world datasets demonstrate that the proposed algorithm outperforms other benchmarks in terms of the learning accuracy and energy consumption.
arXiv Detail & Related papers (2023-08-04T04:18:01Z) - FAStEN: An Efficient Adaptive Method for Feature Selection and Estimation in High-Dimensional Functional Regressions [7.674715791336311]
We propose a new, flexible and ultra-efficient approach to perform feature selection in a sparse function-on-function regression problem.
We show how to extend it to the scalar-on-function framework.
We present an application to brain fMRI data from the AOMIC PIOP1 study.
arXiv Detail & Related papers (2023-03-26T19:41:17Z) - Offline Reinforcement Learning with Differentiable Function
Approximation is Provably Efficient [65.08966446962845]
offline reinforcement learning, which aims at optimizing decision-making strategies with historical data, has been extensively applied in real-life applications.
We take a step by considering offline reinforcement learning with differentiable function class approximation (DFA)
Most importantly, we show offline differentiable function approximation is provably efficient by analyzing the pessimistic fitted Q-learning algorithm.
arXiv Detail & Related papers (2022-10-03T07:59:42Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - On the utility of power spectral techniques with feature selection
techniques for effective mental task classification in noninvasive BCI [19.19039983741124]
This paper proposes an approach to select relevant and non-redundant spectral features for the mental task classification.
The findings demonstrate substantial improvements in the performance of the learning model for mental task classification.
arXiv Detail & Related papers (2021-11-16T00:27:53Z) - RFCBF: enhance the performance and stability of Fast Correlation-Based
Filter [6.781877756322586]
We propose a novel extension of FCBF, called RFCBF, which combines resampling technique to improve classification accuracy.
The experimental results show that the RFCBF algorithm yields significantly better results than previous state-of-the-art methods in terms of classification accuracy and runtime.
arXiv Detail & Related papers (2021-05-30T12:36:32Z) - Feature Selection for Imbalanced Data with Deep Sparse Autoencoders
Ensemble [0.5352699766206808]
Class imbalance is a common issue in many domain applications of learning algorithms.
We propose a filtering FS algorithm ranking feature importance on the basis of the Reconstruction Error of a Deep Sparse AutoEncoders Ensemble.
We empirically demonstrate the efficacy of our algorithm in several experiments on high-dimensional datasets of varying sample size.
arXiv Detail & Related papers (2021-03-22T09:17:08Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z) - Dynamic Federated Learning [57.14673504239551]
Federated learning has emerged as an umbrella term for centralized coordination strategies in multi-agent environments.
We consider a federated learning model where at every iteration, a random subset of available agents perform local updates based on their data.
Under a non-stationary random walk model on the true minimizer for the aggregate optimization problem, we establish that the performance of the architecture is determined by three factors, namely, the data variability at each agent, the model variability across all agents, and a tracking term that is inversely proportional to the learning rate of the algorithm.
arXiv Detail & Related papers (2020-02-20T15:00:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.