Revisiting Classical Multiclass Linear Discriminant Analysis with a
Novel Prototype-based Interpretable Solution
- URL: http://arxiv.org/abs/2205.00668v1
- Date: Mon, 2 May 2022 06:12:42 GMT
- Title: Revisiting Classical Multiclass Linear Discriminant Analysis with a
Novel Prototype-based Interpretable Solution
- Authors: Sayed Kamaledin Ghiasi-Shirazi
- Abstract summary: We introduce a novel solution to classical LDA, called LDA++, that yields $C$ features, each one interpretable as measuring similarity to one cluster.
This novel solution bridges between dimensionality reduction and multiclass classification.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Linear discriminant analysis (LDA) is a fundamental method for feature
extraction and dimensionality reduction. Despite having many variants,
classical LDA has its importance, as it is a keystone in human knowledge about
pattern recognition. For a dataset containing $C$ clusters, the classical
solution to LDA extracts at most $C-1$ features. In this paper, we introduce a
novel solution to classical LDA, called LDA++, that yields $C$ features, each
one interpretable as measuring similarity to one cluster. This novel solution
bridges between dimensionality reduction and multiclass classification.
Specifically, we prove that, under some mild conditions, the optimal weights of
a linear multiclass classifier for homoscedastic Gaussian data also make an
optimal solution to LDA. In addition, this novel interpretable solution reveals
some new facts about LDA and its relation with PCA. We provide a complete
numerical solution for our novel method, covering the cases 1) when the scatter
matrices can be constructed explicitly, 2) when constructing the scatter
matrices is infeasible, and 3) the kernel extension. The code is available at
https://github.com/k-ghiasi/LDA-plus-plus.
Related papers
- Synergistic eigenanalysis of covariance and Hessian matrices for enhanced binary classification [72.77513633290056]
We present a novel approach that combines the eigenanalysis of a covariance matrix evaluated on a training set with a Hessian matrix evaluated on a deep learning model.
Our method captures intricate patterns and relationships, enhancing classification performance.
arXiv Detail & Related papers (2024-02-14T16:10:42Z) - Minimally Informed Linear Discriminant Analysis: training an LDA model
with unlabelled data [51.673443581397954]
We show that it is possible to compute the exact projection vector from LDA models based on unlabelled data.
We show that the MILDA projection vector can be computed in a closed form with a computational cost comparable to LDA.
arXiv Detail & Related papers (2023-10-17T09:50:31Z) - Sparse Linear Centroid-Encoder: A Convex Method for Feature Selection [1.057079240576682]
We present Sparse Centroid-Encoder (SLCE) over a novel feature selection technique.
The algorithm uses a linear network to reconstruct a neural feature at the same time.
arXiv Detail & Related papers (2023-06-07T23:07:55Z) - GO-LDA: Generalised Optimal Linear Discriminant Analysis [6.644357197885522]
Linear discriminant analysis has been a useful tool in pattern recognition and data analysis research and practice.
We show that the generalised eigenanalysis solution to multiclass LDA does neither yield orthogonal discriminant directions nor maximise discrimination of projected data along them.
arXiv Detail & Related papers (2023-05-23T23:11:05Z) - Divide-and-Conquer Hard-thresholding Rules in High-dimensional
Imbalanced Classification [1.0312968200748118]
We study the impact of imbalance class sizes on the linear discriminant analysis (LDA) in high dimensions.
We show that due to data scarcity in one class, referred to as the minority class, the LDA ignores the minority class yielding a maximum misclassification rate.
We propose a new construction of a hard-conquering rule based on a divide-and-conquer technique that reduces the large difference between the misclassification rates.
arXiv Detail & Related papers (2021-11-05T07:44:28Z) - Linear-Sample Learning of Low-Rank Distributions [56.59844655107251]
We show that learning $ktimes k$, rank-$r$, matrices to normalized $L_1$ distance requires $Omega(frackrepsilon2)$ samples.
We propose an algorithm that uses $cal O(frackrepsilon2log2fracepsilon)$ samples, a number linear in the high dimension, and nearly linear in the matrices, typically low, rank proofs.
arXiv Detail & Related papers (2020-09-30T19:10:32Z) - Self-Weighted Robust LDA for Multiclass Classification with Edge Classes [111.5515086563592]
A novel self-weighted robust LDA with l21-norm based between-class distance criterion, called SWRLDA, is proposed for multi-class classification.
The proposed SWRLDA is easy to implement, and converges fast in practice.
arXiv Detail & Related papers (2020-09-24T12:32:55Z) - High-Dimensional Quadratic Discriminant Analysis under Spiked Covariance
Model [101.74172837046382]
We propose a novel quadratic classification technique, the parameters of which are chosen such that the fisher-discriminant ratio is maximized.
Numerical simulations show that the proposed classifier not only outperforms the classical R-QDA for both synthetic and real data but also requires lower computational complexity.
arXiv Detail & Related papers (2020-06-25T12:00:26Z) - Saliency-based Weighted Multi-label Linear Discriminant Analysis [101.12909759844946]
We propose a new variant of Linear Discriminant Analysis (LDA) to solve multi-label classification tasks.
The proposed method is based on a probabilistic model for defining the weights of individual samples.
The Saliency-based weighted Multi-label LDA approach is shown to lead to performance improvements in various multi-label classification problems.
arXiv Detail & Related papers (2020-04-08T19:40:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.