In-situ animal behavior classification using knowledge distillation and
fixed-point quantization
- URL: http://arxiv.org/abs/2209.04130v1
- Date: Fri, 9 Sep 2022 06:07:17 GMT
- Title: In-situ animal behavior classification using knowledge distillation and
fixed-point quantization
- Authors: Reza Arablouei, Liang Wang, Caitlin Phillips, Lachlan Currie, Jordan
Yates, Greg Bishop-Hurley
- Abstract summary: We take a deep and complex convolutional neural network, known as residual neural network (ResNet), as the teacher model.
We implement both unquantized and quantized versions of the developed KD-based models on the embedded systems of our purpose-built collar and ear tag devices.
- Score: 6.649514998517633
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We explore the use of knowledge distillation (KD) for learning compact and
accurate models that enable classification of animal behavior from
accelerometry data on wearable devices. To this end, we take a deep and complex
convolutional neural network, known as residual neural network (ResNet), as the
teacher model. ResNet is specifically designed for multivariate time-series
classification. We use ResNet to distil the knowledge of animal behavior
classification datasets into soft labels, which consist of the predicted
pseudo-probabilities of every class for each datapoint. We then use the soft
labels to train our significantly less complex student models, which are based
on the gated recurrent unit (GRU) and multilayer perceptron (MLP). The
evaluation results using two real-world animal behavior classification datasets
show that the classification accuracy of the student GRU-MLP models improves
appreciably through KD, approaching that of the teacher ResNet model. To
further reduce the computational and memory requirements of performing
inference using the student models trained via KD, we utilize dynamic
fixed-point quantization through an appropriate modification of the
computational graphs of the models. We implement both unquantized and quantized
versions of the developed KD-based models on the embedded systems of our
purpose-built collar and ear tag devices to classify animal behavior in situ
and in real time. The results corroborate the effectiveness of KD and
quantization in improving the inference performance in terms of both
classification accuracy and computational and memory efficiency.
Related papers
- Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals [91.59906995214209]
We propose a new evaluation method, Counterfactual Attentiveness Test (CAT)
CAT uses counterfactuals by replacing part of the input with its counterpart from a different example, expecting an attentive model to change its prediction.
We show that GPT3 becomes less attentive with an increased number of demonstrations, while its accuracy on the test data improves.
arXiv Detail & Related papers (2023-11-16T06:27:35Z) - Directed Acyclic Graph Factorization Machines for CTR Prediction via
Knowledge Distillation [65.62538699160085]
We propose a Directed Acyclic Graph Factorization Machine (KD-DAGFM) to learn the high-order feature interactions from existing complex interaction models for CTR prediction via Knowledge Distillation.
KD-DAGFM achieves the best performance with less than 21.5% FLOPs of the state-of-the-art method on both online and offline experiments.
arXiv Detail & Related papers (2022-11-21T03:09:42Z) - Discover, Explanation, Improvement: An Automatic Slice Detection
Framework for Natural Language Processing [72.14557106085284]
slice detection models (SDM) automatically identify underperforming groups of datapoints.
This paper proposes a benchmark named "Discover, Explain, improve (DEIM)" for classification NLP tasks.
Our evaluation shows that Edisa can accurately select error-prone datapoints with informative semantic features.
arXiv Detail & Related papers (2022-11-08T19:00:00Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - Deep Learning Models for Knowledge Tracing: Review and Empirical
Evaluation [2.423547527175807]
We review and evaluate a body of deep learning knowledge tracing (DLKT) models with openly available and widely-used data sets.
The evaluated DLKT models have been reimplemented for assessing and replicability of previously reported results.
arXiv Detail & Related papers (2021-12-30T14:19:27Z) - Animal Behavior Classification via Accelerometry Data and Recurrent
Neural Networks [11.099308746733028]
We study the classification of animal behavior using accelerometry data through various recurrent neural network (RNN) models.
We evaluate the classification performance and complexity of the considered models.
We also include two state-of-the-art convolutional neural network (CNN)-based time-series classification models in the evaluations.
arXiv Detail & Related papers (2021-11-24T23:28:25Z) - Learning to Rectify for Robust Learning with Noisy Labels [25.149277009932423]
We propose warped probabilistic inference (WarPI) to achieve adaptively rectifying the training procedure for the classification network.
We evaluate WarPI on four benchmarks of robust learning with noisy labels and achieve the new state-of-the-art under variant noise types.
arXiv Detail & Related papers (2021-11-08T02:25:50Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - Efficient training of lightweight neural networks using Online
Self-Acquired Knowledge Distillation [51.66271681532262]
Online Self-Acquired Knowledge Distillation (OSAKD) is proposed, aiming to improve the performance of any deep neural model in an online manner.
We utilize k-nn non-parametric density estimation technique for estimating the unknown probability distributions of the data samples in the output feature space.
arXiv Detail & Related papers (2021-08-26T14:01:04Z) - Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized
Deep Neural Networks [27.533162215182422]
quantization of deep neural networks (QDNNs) has been actively studied for deployment in edge devices.
Recent studies employ the knowledge distillation (KD) method to improve the performance of quantized networks.
In this study, we propose ensemble training for QDNNs (SPEQ)
arXiv Detail & Related papers (2020-09-30T08:38:37Z) - Supervised Learning in the Presence of Concept Drift: A modelling
framework [5.22609266390809]
We present a modelling framework for the investigation of supervised learning in non-stationary environments.
We model two example types of learning systems: prototype-based Learning Vector Quantization (LVQ) for classification and shallow, layered neural networks for regression tasks.
arXiv Detail & Related papers (2020-05-21T09:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.