Combining Primal and Dual Representations in Deep Restricted Kernel
Machines Classifiers
- URL: http://arxiv.org/abs/2306.07015v2
- Date: Tue, 29 Aug 2023 09:17:05 GMT
- Title: Combining Primal and Dual Representations in Deep Restricted Kernel
Machines Classifiers
- Authors: Francesco Tonin, Panagiotis Patrinos, Johan A. K. Suykens
- Abstract summary: We propose a new method for DRKM classification coupling the objectives of KPCA and classification levels.
The classification level can be formulated as an LSSVM or as a primal feature map, combining depth in terms of levels and layers.
We show that our developed algorithm can effectively learn from small datasets, while using less memory than the convolutional neural network (CNN) with high-dimensional data.
- Score: 17.031744210104556
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the context of deep learning with kernel machines, the deep Restricted
Kernel Machine (DRKM) framework allows multiple levels of kernel PCA (KPCA) and
Least-Squares Support Vector Machines (LSSVM) to be combined into a deep
architecture using visible and hidden units. We propose a new method for DRKM
classification coupling the objectives of KPCA and classification levels, with
the hidden feature matrix lying on the Stiefel manifold. The classification
level can be formulated as an LSSVM or as an MLP feature map, combining depth
in terms of levels and layers. The classification level is expressed in its
primal formulation, as the deep KPCA levels, in their dual formulation, can
embed the most informative components of the data in a much lower dimensional
space. The dual setting is independent of the dimension of the inputs and the
primal setting is parametric, which makes the proposed method computationally
efficient for both high-dimensional inputs and large datasets. In the
experiments, we show that our developed algorithm can effectively learn from
small datasets, while using less memory than the convolutional neural network
(CNN) with high-dimensional data. and that models with multiple KPCA levels can
outperform models with a single level. On the tested larger-scale datasets,
DRKM is more energy efficient than CNN while maintaining comparable
performance.
Related papers
- Minimally Supervised Learning using Topological Projections in
Self-Organizing Maps [55.31182147885694]
We introduce a semi-supervised learning approach based on topological projections in self-organizing maps (SOMs)
Our proposed method first trains SOMs on unlabeled data and then a minimal number of available labeled data points are assigned to key best matching units (BMU)
Our results indicate that the proposed minimally supervised model significantly outperforms traditional regression techniques.
arXiv Detail & Related papers (2024-01-12T22:51:48Z) - Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - Qudit Machine Learning [0.0]
We present a comprehensive investigation into the learning capabilities of a simple d-level system (qudit)
Our study is specialized for classification tasks using real-world databases, specifically the Iris, breast cancer, and MNIST datasets.
arXiv Detail & Related papers (2023-08-30T18:00:04Z) - Feature Selection using Sparse Adaptive Bottleneck Centroid-Encoder [1.2487990897680423]
We introduce a novel nonlinear model, Sparse Adaptive Bottleneckid-Encoder (SABCE), for determining the features that discriminate between two or more classes.
The algorithm is applied to various real-world data sets, including high-dimensional biological, image, speech, and accelerometer sensor data.
arXiv Detail & Related papers (2023-06-07T21:37:21Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - A multi-model-based deep learning framework for short text multiclass
classification with the imbalanced and extremely small data set [0.6875312133832077]
This paper proposes a multimodel-based deep learning framework for short-text multiclass classification with an imbalanced and extremely small data set.
It retains the state-of-the-art baseline performance in terms of precision, recall, accuracy, and F1 score.
arXiv Detail & Related papers (2022-06-24T00:51:02Z) - Memory and Computation-Efficient Kernel SVM via Binary Embedding and
Ternary Model Coefficients [18.52747917850984]
Kernel approximation is widely used to scale up kernel SVM training and prediction.
Memory and computation costs of kernel approximation models are still too high if we want to deploy them on memory-limited devices.
We propose a novel memory and computation-efficient kernel SVM model by using both binary embedding and binary model coefficients.
arXiv Detail & Related papers (2020-10-06T09:41:54Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - A Generic Network Compression Framework for Sequential Recommender
Systems [71.81962915192022]
Sequential recommender systems (SRS) have become the key technology in capturing user's dynamic interests and generating high-quality recommendations.
We propose a compressed sequential recommendation framework, termed as CpRec, where two generic model shrinking techniques are employed.
By the extensive ablation studies, we demonstrate that the proposed CpRec can achieve up to 4$sim$8 times compression rates in real-world SRS datasets.
arXiv Detail & Related papers (2020-04-21T08:40:55Z) - On Coresets for Support Vector Machines [61.928187390362176]
A coreset is a small, representative subset of the original data points.
We show that our algorithm can be used to extend the applicability of any off-the-shelf SVM solver to streaming, distributed, and dynamic data settings.
arXiv Detail & Related papers (2020-02-15T23:25:12Z) - Efficient Structure-preserving Support Tensor Train Machine [0.0]
Train Multi-way Multi-level Kernel (TT-MMK)
We develop the Train Multi-way Multi-level Kernel (TT-MMK), which combines the simplicity of the Polyadic decomposition, the classification power of the Dual Structure-preserving Support Machine, and the reliability of the Train Vector approximation.
We show by experiments that the TT-MMK method is usually more reliable, less sensitive to tuning parameters, and gives higher prediction accuracy in the SVM classification when benchmarked against other state-of-the-art techniques.
arXiv Detail & Related papers (2020-02-12T16:35:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.