Weight Vector Tuning and Asymptotic Analysis of Binary Linear
Classifiers
- URL: http://arxiv.org/abs/2110.00567v1
- Date: Fri, 1 Oct 2021 17:50:46 GMT
- Title: Weight Vector Tuning and Asymptotic Analysis of Binary Linear
Classifiers
- Authors: Lama B. Niyazi, Abla Kammoun, Hayssam Dahrouj, Mohamed-Slim Alouini,
and Tareq Al-Naffouri
- Abstract summary: This paper proposes weight vector tuning of a generic binary linear classifier through the parameterization of a decomposition of the discriminant by a scalar.
It is also found that weight vector tuning significantly improves the performance of Linear Discriminant Analysis (LDA) under high estimation noise.
- Score: 82.5915112474988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unlike its intercept, a linear classifier's weight vector cannot be tuned by
a simple grid search. Hence, this paper proposes weight vector tuning of a
generic binary linear classifier through the parameterization of a
decomposition of the discriminant by a scalar which controls the trade-off
between conflicting informative and noisy terms. By varying this parameter, the
original weight vector is modified in a meaningful way. Applying this method to
a number of linear classifiers under a variety of data dimensionality and
sample size settings reveals that the classification performance loss due to
non-optimal native hyperparameters can be compensated for by weight vector
tuning. This yields computational savings as the proposed tuning method reduces
to tuning a scalar compared to tuning the native hyperparameter, which may
involve repeated weight vector generation along with its burden of
optimization, dimensionality reduction, etc., depending on the classifier. It
is also found that weight vector tuning significantly improves the performance
of Linear Discriminant Analysis (LDA) under high estimation noise. Proceeding
from this second finding, an asymptotic study of the misclassification
probability of the parameterized LDA classifier in the growth regime where the
data dimensionality and sample size are comparable is conducted. Using random
matrix theory, the misclassification probability is shown to converge to a
quantity that is a function of the true statistics of the data. Additionally,
an estimator of the misclassification probability is derived. Finally,
computationally efficient tuning of the parameter using this estimator is
demonstrated on real data.
Related papers
- Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Regularized Linear Discriminant Analysis Using a Nonlinear Covariance
Matrix Estimator [11.887333567383239]
Linear discriminant analysis (LDA) is a widely used technique for data classification.
LDA becomes inefficient when the data covariance matrix is ill-conditioned.
Regularized LDA methods have been proposed to cope with such a situation.
arXiv Detail & Related papers (2024-01-31T11:37:14Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - A Doubly Regularized Linear Discriminant Analysis Classifier with
Automatic Parameter Selection [24.027886914804775]
Linear discriminant analysis (LDA) based classifiers tend to falter in many practical settings where the training data size is smaller than, or comparable to, the number of features.
We propose a doubly regularized LDA classifier that we denote as R2LDA.
Results obtained from both synthetic and real data demonstrate the consistency and effectiveness of the proposed R2LDA approach.
arXiv Detail & Related papers (2020-04-28T07:09:22Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - A working likelihood approach to support vector regression with a
data-driven insensitivity parameter [2.842794675894731]
The insensitive parameter in support vector regression determines the set of support vectors that greatly impacts the prediction.
A data-driven approach is proposed to determine an approximate value for this insensitive parameter.
This data-driven support vector regression also statistically standardizes samples using the scale of noises.
arXiv Detail & Related papers (2020-03-09T02:32:32Z) - Implicit differentiation of Lasso-type models for hyperparameter
optimization [82.73138686390514]
We introduce an efficient implicit differentiation algorithm, without matrix inversion, tailored for Lasso-type problems.
Our approach scales to high-dimensional data by leveraging the sparsity of the solutions.
arXiv Detail & Related papers (2020-02-20T18:43:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.