A connection between the pattern classification problem and the General
Linear Model for statistical inference
- URL: http://arxiv.org/abs/2012.08903v1
- Date: Wed, 16 Dec 2020 12:26:26 GMT
- Title: A connection between the pattern classification problem and the General
Linear Model for statistical inference
- Authors: Juan Manuel Gorriz and SIPBA group and John Suckling
- Abstract summary: Both approaches, i.e. GLM and LRM, apply to different domains, the observation and the label domains.
We derive a statistical test based on a more refined predictive algorithm.
The MLE-based inference employs a residual score and includes the upper bound to compute a better estimation of the actual (real) error.
- Score: 0.2320417845168326
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A connection between the General Linear Model (GLM) in combination with
classical statistical inference and the machine learning (MLE)-based inference
is described in this paper. Firstly, the estimation of the GLM parameters is
expressed as a Linear Regression Model (LRM) of an indicator matrix, that is,
in terms of the inverse problem of regressing the observations. In other words,
both approaches, i.e. GLM and LRM, apply to different domains, the observation
and the label domains, and are linked by a normalization value at the
least-squares solution. Subsequently, from this relationship we derive a
statistical test based on a more refined predictive algorithm, i.e. the
(non)linear Support Vector Machine (SVM) that maximizes the class margin of
separation, within a permutation analysis. The MLE-based inference employs a
residual score and includes the upper bound to compute a better estimation of
the actual (real) error. Experimental results demonstrate how the parameter
estimations derived from each model resulted in different classification
performances in the equivalent inverse problem. Moreover, using real data the
aforementioned predictive algorithms within permutation tests, including such
model-free estimators, are able to provide a good trade-off between type I
error and statistical power.
Related papers
- Agnostic Learning of Mixed Linear Regressions with EM and AM Algorithms [22.79595679373698]
Mixed linear regression is a well-studied problem in statistics and machine learning.
In this paper, we consider the more general problem of learning of mixed linear regression from samples.
We show that the AM and EM algorithms lead to learning in mixed linear regression by converging to the population loss minimizers.
arXiv Detail & Related papers (2024-06-03T09:43:24Z) - Weakly supervised covariance matrices alignment through Stiefel matrices
estimation for MEG applications [64.20396555814513]
This paper introduces a novel domain adaptation technique for time series data, called Mixing model Stiefel Adaptation (MSA)
We exploit abundant unlabeled data in the target domain to ensure effective prediction by establishing pairwise correspondence with equivalent signal variances between domains.
MSA outperforms recent methods in brain-age regression with task variations using magnetoencephalography (MEG) signals from the Cam-CAN dataset.
arXiv Detail & Related papers (2024-01-24T19:04:49Z) - Global Convergence of Online Identification for Mixed Linear Regression [1.9295130374196499]
Mixed linear regression (MLR) is a powerful model for characterizing nonlinear relationships.
This paper investigates the online identification and data clustering problems for two basic classes of MLRs.
It introduces two corresponding new online identification algorithms based on the expectation-maximization principle.
arXiv Detail & Related papers (2023-11-30T12:30:42Z) - Generalized generalized linear models: Convex estimation and online
bounds [11.295523372922533]
We introduce inequalities in a class of models (GL-based) models (GGLM)
The proposed approach uses the operator-based approach to overcome nontemporal variation among models.
We demonstrate the performance using numerical simulations and a real data example for incidents.
arXiv Detail & Related papers (2023-04-26T19:19:42Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Robust Regularized Low-Rank Matrix Models for Regression and
Classification [14.698622796774634]
We propose a framework for matrix variate regression models based on a rank constraint, vector regularization (e.g., sparsity), and a general loss function.
We show that the algorithm is guaranteed to converge, all accumulation points of the algorithm have estimation errors in the order of $O(sqrtn)$ally and substantially attaining the minimax rate.
arXiv Detail & Related papers (2022-05-14T18:03:48Z) - A hypothesis-driven method based on machine learning for neuroimaging
data analysis [0.0]
Machine learning approaches for discrimination of spatial patterns of brain images have limited their operation to feature extraction and linear classification tasks.
We show that the estimation of the conventional General linear Model (GLM) has been connected to an univariate classification task.
We derive a refined statistical test with the GLM based on the parameters obtained by a linear Support Vector Regression (SVR) in the emphinverse problem (SVR-iGLM)
Using real data from a multisite initiative the proposed MLE-based inference demonstrates statistical power and the control of false positives, outperforming the regular G
arXiv Detail & Related papers (2022-02-09T11:13:02Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.