Audio Tagging on an Embedded Hardware Platform
- URL: http://arxiv.org/abs/2306.09106v1
- Date: Thu, 15 Jun 2023 13:02:41 GMT
- Title: Audio Tagging on an Embedded Hardware Platform
- Authors: Gabriel Bibbo, Arshdeep Singh, Mark D. Plumbley
- Abstract summary: We analyze how the performance of large-scale pretrained audio neural networks changes when deployed on a hardware such as Raspberry Pi.
Our experiments reveal that the continuous CPU usage results in an increased temperature that can trigger an automated slowdown mechanism.
The quality of a microphone, specifically with affordable devices like the Google AIY Voice Kit, and audio signal volume, all affect the system performance.
- Score: 20.028643659869573
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks (CNNs) have exhibited state-of-the-art
performance in various audio classification tasks. However, their real-time
deployment remains a challenge on resource-constrained devices like embedded
systems. In this paper, we analyze how the performance of large-scale
pretrained audio neural networks designed for audio pattern recognition changes
when deployed on a hardware such as Raspberry Pi. We empirically study the role
of CPU temperature, microphone quality and audio signal volume on performance.
Our experiments reveal that the continuous CPU usage results in an increased
temperature that can trigger an automated slowdown mechanism in the Raspberry
Pi, impacting inference latency. The quality of a microphone, specifically with
affordable devices like the Google AIY Voice Kit, and audio signal volume, all
affect the system performance. In the course of our investigation, we encounter
substantial complications linked to library compatibility and the unique
processor architecture requirements of the Raspberry Pi, making the process
less straightforward compared to conventional computers (PCs). Our
observations, while presenting challenges, pave the way for future researchers
to develop more compact machine learning models, design heat-dissipative
hardware, and select appropriate microphones when AI models are deployed for
real-time applications on edge devices. All related assets and an interactive
demo can be found on GitHub
Related papers
- Neurobench: DCASE 2020 Acoustic Scene Classification benchmark on XyloAudio 2 [0.06752396542927405]
XyloAudio is a line of ultra-low-power audio inference chips.
It is designed for in- and near-microphone analysis of audio in real-time energy-constrained scenarios.
arXiv Detail & Related papers (2024-10-31T09:48:12Z) - MAPLE-X: Latency Prediction with Explicit Microprocessor Prior Knowledge [87.41163540910854]
Deep neural network (DNN) latency characterization is a time-consuming process.
We propose MAPLE-X which extends MAPLE by incorporating explicit prior knowledge of hardware devices and DNN architecture latency.
arXiv Detail & Related papers (2022-05-25T11:08:20Z) - A Study of Designing Compact Audio-Visual Wake Word Spotting System
Based on Iterative Fine-Tuning in Neural Network Pruning [57.28467469709369]
We investigate on designing a compact audio-visual wake word spotting (WWS) system by utilizing visual information.
We introduce a neural network pruning strategy via the lottery ticket hypothesis in an iterative fine-tuning manner (LTH-IF)
The proposed audio-visual system achieves significant performance improvements over the single-modality (audio-only or video-only) system under different noisy conditions.
arXiv Detail & Related papers (2022-02-17T08:26:25Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - MAPLE: Microprocessor A Priori for Latency Estimation [81.91509153539566]
Modern deep neural networks must demonstrate state-of-the-art accuracy while exhibiting low latency and energy consumption.
Measuring the latency of every evaluated architecture adds a significant amount of time to the NAS process.
We propose Microprocessor A Priori for Estimation Estimation MAPLE that does not rely on transfer learning or domain adaptation.
arXiv Detail & Related papers (2021-11-30T03:52:15Z) - DeepSpectrumLite: A Power-Efficient Transfer Learning Framework for
Embedded Speech and Audio Processing from Decentralised Data [0.0]
We introduce DeepSpectrumLite, an open-source, lightweight transfer learning framework for on-device speech and audio recognition.
The framework creates and augments Mel-spectrogram plots on-the-fly from raw audio signals which are then used to finetune specific pre-trained CNNs.
The whole pipeline can be run in real-time with a mean inference lag of 242.0 ms when a DenseNet121 model is used on a consumer-grade Motorola moto e7 plus smartphone.
arXiv Detail & Related papers (2021-04-23T14:32:33Z) - Dissecting User-Perceived Latency of On-Device E2E Speech Recognition [34.645194215436966]
We show that factors affecting token emission latency, and endpointing behavior significantly impact on user-perceived latency (UPL)
We achieve the best trade-off between latency and word error rate when performing ASR jointly with endpointing, and using the recently proposed alignment regularization.
arXiv Detail & Related papers (2021-04-06T00:55:11Z) - Real-time Timbre Transfer and Sound Synthesis using DDSP [1.7942265700058984]
We present a real-time implementation of the MagentaP library embedded in a virtual synthesizer as a plug-in.
We focused on timbre transfer from learned representations of real instruments to arbitrary sound inputs as well as controlling these models by MIDI.
We developed a GUI for intuitive high-level controls which can be used for post-processing and manipulating the parameters estimated by the neural network.
arXiv Detail & Related papers (2021-03-12T11:49:51Z) - LightSpeech: Lightweight and Fast Text to Speech with Neural
Architecture Search [127.56834100382878]
We propose LightSpeech to automatically design more lightweight and efficient TTS models based on FastSpeech.
Experiments show that the model discovered by our method achieves 15x model compression ratio and 6.5x inference speedup on CPU with on par voice quality.
arXiv Detail & Related papers (2021-02-08T07:45:06Z) - Deep Speaker Embeddings for Far-Field Speaker Recognition on Short
Utterances [53.063441357826484]
Speaker recognition systems based on deep speaker embeddings have achieved significant performance in controlled conditions.
Speaker verification on short utterances in uncontrolled noisy environment conditions is one of the most challenging and highly demanded tasks.
This paper presents approaches aimed to achieve two goals: a) improve the quality of far-field speaker verification systems in the presence of environmental noise, reverberation and b) reduce the system qualitydegradation for short utterances.
arXiv Detail & Related papers (2020-02-14T13:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.