Ultra-compact Binary Neural Networks for Human Activity Recognition on
RISC-V Processors
- URL: http://arxiv.org/abs/2205.12781v1
- Date: Wed, 25 May 2022 13:52:35 GMT
- Title: Ultra-compact Binary Neural Networks for Human Activity Recognition on
RISC-V Processors
- Authors: Francesco Daghero, Chen Xie, Daniele Jahier Pagliari, Alessio
Burrello, Marco Castellano, Luca Gandolfi, Andrea Calimera, Enrico Macii,
Massimo Poncino
- Abstract summary: Human Activity Recognition (HAR) is a relevant inference task in many mobile applications.
We propose a novel implementation of HAR based on deep neural networks, and precisely on Binary Neural Networks (BNNs)
BNNs yield very small memory footprints and low inference complexity, thanks to the replacement of arithmetic operations with bit-wise ones.
- Score: 10.195581493173643
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human Activity Recognition (HAR) is a relevant inference task in many mobile
applications. State-of-the-art HAR at the edge is typically achieved with
lightweight machine learning models such as decision trees and Random Forests
(RFs), whereas deep learning is less common due to its high computational
complexity. In this work, we propose a novel implementation of HAR based on
deep neural networks, and precisely on Binary Neural Networks (BNNs), targeting
low-power general purpose processors with a RISC-V instruction set. BNNs yield
very small memory footprints and low inference complexity, thanks to the
replacement of arithmetic operations with bit-wise ones. However, existing BNN
implementations on general purpose processors impose constraints tailored to
complex computer vision tasks, which result in over-parametrized models for
simpler problems like HAR. Therefore, we also introduce a new BNN inference
library, which targets ultra-compact models explicitly. With experiments on a
single-core RISC-V processor, we show that BNNs trained on two HAR datasets
obtain higher classification accuracy compared to a state-of-the-art baseline
based on RFs. Furthermore, our BNN reaches the same accuracy of a RF with
either less memory (up to 91%) or more energy-efficiency (up to 70%), depending
on the complexity of the features extracted by the RF.
Related papers
- Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Low-bit Shift Network for End-to-End Spoken Language Understanding [7.851607739211987]
We propose the use of power-of-two quantization, which quantizes continuous parameters into low-bit power-of-two values.
This reduces computational complexity by removing expensive multiplication operations and with the use of low-bit weights.
arXiv Detail & Related papers (2022-07-15T14:34:22Z) - A Robust Deep Learning-Based Beamforming Design for RIS-assisted
Multiuser MISO Communications with Practical Constraints [4.727307803726522]
We consider a RIS-aided multiuser multiple-input single-output downlink communication system.
We develop a deep quantization neural network (DQNN) to simultaneously design the active and passive beamforming.
We extend the two proposed DQNN-based algorithms to the case that the discrete phase shifts and imperfect CSI are considered simultaneously.
arXiv Detail & Related papers (2021-11-12T03:53:20Z) - Binary Complex Neural Network Acceleration on FPGA [19.38270650475235]
Binarized Complex Neural Network (BCNN) shows great potential in classifying complex data in real-time.
We propose a structural pruning based accelerator of BCNN, which is able to provide more than 5000 frames/s inference throughput on edge devices.
arXiv Detail & Related papers (2021-08-10T17:53:30Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - NL-CNN: A Resources-Constrained Deep Learning Model based on Nonlinear
Convolution [0.0]
A novel convolution neural network model, abbreviated NL-CNN, is proposed, where nonlinear convolution is emulated in a cascade of convolution + nonlinearity layers.
Performance evaluation for several widely known datasets is provided, showing several relevant features.
arXiv Detail & Related papers (2021-01-30T13:38:42Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Self-Organized Operational Neural Networks with Generative Neurons [87.32169414230822]
ONNs are heterogenous networks with a generalized neuron model that can encapsulate any set of non-linear operators.
We propose Self-organized ONNs (Self-ONNs) with generative neurons that have the ability to adapt (optimize) the nodal operator of each connection.
arXiv Detail & Related papers (2020-04-24T14:37:56Z) - Constructing Deep Neural Networks with a Priori Knowledge of Wireless
Tasks [37.060397377445504]
Two kinds of permutation invariant properties widely existed in wireless tasks can be harnessed to reduce the number of model parameters.
We find special architecture of DNNs whose input-output relationships satisfy the properties, called permutation invariant DNN (PINN)
We take predictive resource allocation and interference coordination as examples to show how the PINNs can be employed for learning the optimal policy with unsupervised and supervised learning.
arXiv Detail & Related papers (2020-01-29T08:54:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.