An Advantage Using Feature Selection with a Quantum Annealer
- URL: http://arxiv.org/abs/2211.09756v4
- Date: Thu, 1 Jun 2023 14:05:34 GMT
- Title: An Advantage Using Feature Selection with a Quantum Annealer
- Authors: Andrew Vlasic, Hunter Grant and Salvatore Certo
- Abstract summary: Feature selection is a technique in statistical prediction modeling that identifies features in a record with a strong statistical connection to the target variable.
This paper tests this intuition against classical methods by utilizing open-source data sets and evaluate the efficacy of each trained statistical model.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feature selection is a technique in statistical prediction modeling that
identifies features in a record with a strong statistical connection to the
target variable. Excluding features with a weak statistical connection to the
target variable in training not only drops the dimension of the data, which
decreases the time complexity of the algorithm, it also decreases noise within
the data which assists in avoiding overfitting. In all, feature selection
assists in training a robust statistical model that performs well and is
stable. Given the lack of scalability in classical computation, current
techniques only consider the predictive power of the feature and not redundancy
between the features themselves. Recent advancements in feature selection that
leverages quantum annealing (QA) gives a scalable technique that aims to
maximize the predictive power of the features while minimizing redundancy. As a
consequence, it is expected that this algorithm would assist in the
bias/variance trade-off yielding better features for training a statistical
model. This paper tests this intuition against classical methods by utilizing
open-source data sets and evaluate the efficacy of each trained statistical
model well-known prediction algorithms. The numerical results display an
advantage utilizing the features selected from the algorithm that leveraged QA.
Related papers
- Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - FAStEN: An Efficient Adaptive Method for Feature Selection and Estimation in High-Dimensional Functional Regressions [7.674715791336311]
We propose a new, flexible and ultra-efficient approach to perform feature selection in a sparse function-on-function regression problem.
We show how to extend it to the scalar-on-function framework.
We present an application to brain fMRI data from the AOMIC PIOP1 study.
arXiv Detail & Related papers (2023-03-26T19:41:17Z) - RF+clust for Leave-One-Problem-Out Performance Prediction [0.9281671380673306]
We study leave-one-problem-out (LOPO) performance prediction.
We analyze whether standard random forest (RF) model predictions can be improved by calibrating them with a weighted average of performance values.
arXiv Detail & Related papers (2023-01-23T16:14:59Z) - Sequential Attention for Feature Selection [12.89764845700709]
We propose a feature selection algorithm called Sequential Attention that achieves state-of-the-art empirical results for neural networks.
We give theoretical insights into our algorithm for linear regression by showing that an adaptation to this setting is equivalent to the classical Orthogonal Matching Pursuit (OMP) algorithm.
arXiv Detail & Related papers (2022-09-29T15:49:06Z) - Uncertainty Modeling for Out-of-Distribution Generalization [56.957731893992495]
We argue that the feature statistics can be properly manipulated to improve the generalization ability of deep learning models.
Common methods often consider the feature statistics as deterministic values measured from the learned features.
We improve the network generalization ability by modeling the uncertainty of domain shifts with synthesized feature statistics during training.
arXiv Detail & Related papers (2022-02-08T16:09:12Z) - Quantum Annealing for Automated Feature Selection in Stress Detection [1.8047694351309205]
We present a novel methodology for automated feature subset selection from a pool of physiological signals using Quantum Annealing (QA)
Features are extracted from four signal sources: foot EDA, hand EDA, ECG, and respiration.
Results indicate that QA-based feature subset selection performed equally as that of classical techniques.
arXiv Detail & Related papers (2021-06-09T15:17:48Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - Feature Selection Using Reinforcement Learning [0.0]
The space of variables or features that can be used to characterize a particular predictor of interest continues to grow exponentially.
Identifying the most characterizing features that minimizes the variance without jeopardizing the bias of our models is critical to successfully training a machine learning model.
arXiv Detail & Related papers (2021-01-23T09:24:37Z) - Instability, Computational Efficiency and Statistical Accuracy [101.32305022521024]
We develop a framework that yields statistical accuracy based on interplay between the deterministic convergence rate of the algorithm at the population level, and its degree of (instability) when applied to an empirical object based on $n$ samples.
We provide applications of our general results to several concrete classes of models, including Gaussian mixture estimation, non-linear regression models, and informative non-response models.
arXiv Detail & Related papers (2020-05-22T22:30:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.