SynergicLearning: Neural Network-Based Feature Extraction for
Highly-Accurate Hyperdimensional Learning
- URL: http://arxiv.org/abs/2007.15222v2
- Date: Tue, 4 Aug 2020 17:13:37 GMT
- Title: SynergicLearning: Neural Network-Based Feature Extraction for
Highly-Accurate Hyperdimensional Learning
- Authors: Mahdi Nazemi, Amirhossein Esmaili, Arash Fayyazi, Massoud Pedram
- Abstract summary: Machine learning models differ in terms of accuracy, computational/memory complexity, training time, and adaptability.
This work presents a hybrid, synergic machine learning model that excels at all the said characteristics and is suitable for incremental, on-line learning on a chip.
The proposed model has the same level of accuracy (i.e. $pm$1%) as NNs while achieving at least 10% improvement in accuracy compared to HD learning models.
- Score: 3.5024680868164437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning models differ in terms of accuracy, computational/memory
complexity, training time, and adaptability among other characteristics. For
example, neural networks (NNs) are well-known for their high accuracy due to
the quality of their automatic feature extraction while brain-inspired
hyperdimensional (HD) learning models are famous for their quick training,
computational efficiency, and adaptability. This work presents a hybrid,
synergic machine learning model that excels at all the said characteristics and
is suitable for incremental, on-line learning on a chip. The proposed model
comprises an NN and a classifier. The NN acts as a feature extractor and is
specifically trained to work well with the classifier that employs the HD
computing framework. This work also presents a parameterized hardware
implementation of the said feature extraction and classification components
while introducing a compiler that maps any arbitrary NN and/or classifier to
the aforementioned hardware. The proposed hybrid machine learning model has the
same level of accuracy (i.e. $\pm$1%) as NNs while achieving at least 10%
improvement in accuracy compared to HD learning models. Additionally, the
end-to-end hardware realization of the hybrid model improves power efficiency
by 1.60x compared to state-of-the-art, high-performance HD learning
implementations while improving latency by 2.13x. These results have profound
implications for the application of such synergic models in challenging
cognitive tasks.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Spiking Neural Networks in Vertical Federated Learning: Performance Trade-offs [2.1756721838833797]
Federated machine learning enables model training across multiple clients.
Vertical Federated Learning (VFL) deals with instances where the clients have different feature sets of the same samples.
Spiking Neural Networks (SNNs) are being leveraged to enable fast and accurate processing at the edge.
arXiv Detail & Related papers (2024-07-24T23:31:02Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Directed Acyclic Graph Factorization Machines for CTR Prediction via
Knowledge Distillation [65.62538699160085]
We propose a Directed Acyclic Graph Factorization Machine (KD-DAGFM) to learn the high-order feature interactions from existing complex interaction models for CTR prediction via Knowledge Distillation.
KD-DAGFM achieves the best performance with less than 21.5% FLOPs of the state-of-the-art method on both online and offline experiments.
arXiv Detail & Related papers (2022-11-21T03:09:42Z) - Lightweight Hybrid CNN-ELM Model for Multi-building and Multi-floor
Classification [6.154022105385209]
We propose a lightweight combination of CNN and ELM, which provides a quick and accurate classification of building and floor.
As a result, the proposed model is 58% faster than the benchmark, with a slight improvement in the classification accuracy.
arXiv Detail & Related papers (2022-04-21T21:48:01Z) - LCS: Learning Compressible Subspaces for Adaptive Network Compression at
Inference Time [57.52251547365967]
We propose a method for training a "compressible subspace" of neural networks that contains a fine-grained spectrum of models.
We present results for achieving arbitrarily fine-grained accuracy-efficiency trade-offs at inference time for structured and unstructured sparsity.
Our algorithm extends to quantization at variable bit widths, achieving accuracy on par with individually trained networks.
arXiv Detail & Related papers (2021-10-08T17:03:34Z) - NL-CNN: A Resources-Constrained Deep Learning Model based on Nonlinear
Convolution [0.0]
A novel convolution neural network model, abbreviated NL-CNN, is proposed, where nonlinear convolution is emulated in a cascade of convolution + nonlinearity layers.
Performance evaluation for several widely known datasets is provided, showing several relevant features.
arXiv Detail & Related papers (2021-01-30T13:38:42Z) - Superiorities of Deep Extreme Learning Machines against Convolutional
Neural Networks [3.04585143845864]
Deep Learning (DL) is a machine learning procedure for artificial intelligence that analyzes the input data in detail.
DL has a popularity with the common improvements on the graphical processing unit capabilities.
Deep Extreme Learning machines (Deep ELM) is one of the fastest and effective way to meet fast classification problems.
arXiv Detail & Related papers (2021-01-21T08:22:18Z) - Training Deep Neural Networks with Constrained Learning Parameters [4.917317902787792]
A significant portion of deep learning tasks would run on edge computing systems.
We propose the Combinatorial Neural Network Training Algorithm (CoNNTrA)
CoNNTrA trains deep learning models with ternary learning parameters on the MNIST, Iris and ImageNet data sets.
Our results indicate that CoNNTrA models use 32x less memory and have errors at par with the Backpropagation models.
arXiv Detail & Related papers (2020-09-01T16:20:11Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z) - An Image Enhancing Pattern-based Sparsity for Real-time Inference on
Mobile Devices [58.62801151916888]
We introduce a new sparsity dimension, namely pattern-based sparsity that comprises pattern and connectivity sparsity, and becoming both highly accurate and hardware friendly.
Our approach on the new pattern-based sparsity naturally fits into compiler optimization for highly efficient DNN execution on mobile platforms.
arXiv Detail & Related papers (2020-01-20T16:17:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.