Multilevel orthogonal Bochner function subspaces with applications to
robust machine learning
- URL: http://arxiv.org/abs/2110.01729v4
- Date: Fri, 25 Aug 2023 18:23:26 GMT
- Title: Multilevel orthogonal Bochner function subspaces with applications to
robust machine learning
- Authors: Julio Enrique Castrillon-Candas, Dingning Liu, Sicheng Yang, Mark Kon
- Abstract summary: We consider the data as instances of a random field within a relevant Bochner space.
Our key observation is that the classes can predominantly reside in two distinct subspaces.
- Score: 1.533771872970755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In our approach, we consider the data as instances of a random field within a
relevant Bochner space. Our key observation is that the classes can
predominantly reside in two distinct subspaces. To uncover the separation
between these classes, we employ the Karhunen-Loeve expansion and construct the
appropriate subspaces. This allows us to effectively reveal the distinction
between the classes. The novel features forming the above bases are constructed
by applying a coordinate transformation based on the recent Functional Data
Analysis theory for anomaly detection. The associated signal decomposition is
an exact hierarchical tensor product expansion with known optimality properties
for approximating stochastic processes (random fields) with finite dimensional
function spaces. Using a hierarchical finite dimensional expansion of the
nominal class, a series of orthogonal nested subspaces is constructed for
detecting anomalous signal components. Projection coefficients of input data in
these subspaces are then used to train a Machine Learning (ML classifier.
However, due to the split of the signal into nominal and anomalous projection
components, clearer separation surfaces for the classes arise. In fact we show
that with a sufficiently accurate estimation of the covariance structure of the
nominal class, a sharp classification can be obtained. This is particularly
advantageous for large unbalanced datasets. We demonstrate it on a number of
high-dimensional datasets. This approach yields significant increases in
accuracy of ML methods compared to using the same ML algorithm with the
original feature data. Our tests on the Alzheimer's Disease ADNI dataset shows
a dramatic increase in accuracy (from 48% to 89% accuracy). Furthermore, tests
using unbalanced semi-synthetic datasets created from the benchmark GCM dataset
confirm increased accuracy as the dataset becomes more unbalanced.
Related papers
- Symmetry Discovery for Different Data Types [52.2614860099811]
Equivariant neural networks incorporate symmetries into their architecture, achieving higher generalization performance.
We propose LieSD, a method for discovering symmetries via trained neural networks which approximate the input-output mappings of the tasks.
We validate the performance of LieSD on tasks with symmetries such as the two-body problem, the moment of inertia matrix prediction, and top quark tagging.
arXiv Detail & Related papers (2024-10-13T13:39:39Z) - Hodge-Aware Contrastive Learning [101.56637264703058]
Simplicial complexes prove effective in modeling data with multiway dependencies.
We develop a contrastive self-supervised learning approach for processing simplicial data.
arXiv Detail & Related papers (2023-09-14T00:40:07Z) - Random Smoothing Regularization in Kernel Gradient Descent Learning [24.383121157277007]
We present a framework for random smoothing regularization that can adaptively learn a wide range of ground truth functions belonging to the classical Sobolev spaces.
Our estimator can adapt to the structural assumptions of the underlying data and avoid the curse of dimensionality.
arXiv Detail & Related papers (2023-05-05T13:37:34Z) - Sharp-SSL: Selective high-dimensional axis-aligned random projections
for semi-supervised learning [16.673022545571566]
We propose a new method for high-dimensional semi-supervised learning problems.
It is based on the careful aggregation of the results of a low-dimensional procedure applied to many axis-aligned random projections of the data.
arXiv Detail & Related papers (2023-04-18T17:49:02Z) - Learning to Bound Counterfactual Inference in Structural Causal Models
from Observational and Randomised Data [64.96984404868411]
We derive a likelihood characterisation for the overall data that leads us to extend a previous EM-based algorithm.
The new algorithm learns to approximate the (unidentifiability) region of model parameters from such mixed data sources.
It delivers interval approximations to counterfactual results, which collapse to points in the identifiable case.
arXiv Detail & Related papers (2022-12-06T12:42:11Z) - Subspace clustering in high-dimensions: Phase transitions \&
Statistical-to-Computational gap [24.073221004661427]
A simple model to study subspace clustering is the high-dimensional $k$-Gaussian mixture model.
We provide an exact characterization of the statistically optimal reconstruction error in this model in the high-dimensional regime with extensive sparsity.
arXiv Detail & Related papers (2022-05-26T17:47:35Z) - Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic
Uncertainty [58.144520501201995]
Bi-Lipschitz regularization of neural network layers preserve relative distances between data instances in the feature spaces of each layer.
With the use of an attentive set encoder, we propose to meta learn either diagonal or diagonal plus low-rank factors to efficiently construct task specific covariance matrices.
We also propose an inference procedure which utilizes scaled energy to achieve a final predictive distribution.
arXiv Detail & Related papers (2021-10-12T22:04:19Z) - Tensor Laplacian Regularized Low-Rank Representation for Non-uniformly
Distributed Data Subspace Clustering [2.578242050187029]
Low-Rank Representation (LRR) suffers from discarding the locality information of data points in subspace clustering.
We propose a hypergraph model to facilitate having a variable number of adjacent nodes and incorporating the locality information of the data.
Experiments on artificial and real datasets demonstrate the higher accuracy and precision of the proposed method.
arXiv Detail & Related papers (2021-03-06T08:22:24Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.