A character representation enhanced on-device Intent Classification
- URL: http://arxiv.org/abs/2101.04456v1
- Date: Tue, 12 Jan 2021 13:02:05 GMT
- Title: A character representation enhanced on-device Intent Classification
- Authors: Sudeep Deepak Shivnikar, Himanshu Arora, Harichandana B S S
- Abstract summary: We present a novel light-weight architecture for intent classification that can run efficiently on a device.
Our experiments prove that our proposed model outperforms existing approaches and achieves state-of-the-art results on benchmark datasets.
Our model has tiny memory footprint of 5 MB and low inference time of 2 milliseconds, which proves its efficiency in a resource-constrained environment.
- Score: 2.2625832119364153
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Intent classification is an important task in natural language understanding
systems. Existing approaches have achieved perfect scores on the benchmark
datasets. However they are not suitable for deployment on low-resource devices
like mobiles, tablets, etc. due to their massive model size. Therefore, in this
paper, we present a novel light-weight architecture for intent classification
that can run efficiently on a device. We use character features to enrich the
word representation. Our experiments prove that our proposed model outperforms
existing approaches and achieves state-of-the-art results on benchmark
datasets. We also report that our model has tiny memory footprint of ~5 MB and
low inference time of ~2 milliseconds, which proves its efficiency in a
resource-constrained environment.
Related papers
- Adaptable Embeddings Network (AEN) [49.1574468325115]
We introduce Adaptable Embeddings Networks (AEN), a novel dual-encoder architecture using Kernel Density Estimation (KDE)
AEN allows for runtime adaptation of classification criteria without retraining and is non-autoregressive.
The architecture's ability to preprocess and cache condition embeddings makes it ideal for edge computing applications and real-time monitoring systems.
arXiv Detail & Related papers (2024-11-21T02:15:52Z) - EvalCrafter: Benchmarking and Evaluating Large Video Generation Models [70.19437817951673]
We argue that it is hard to judge the large conditional generative models from the simple metrics since these models are often trained on very large datasets with multi-aspect abilities.
Our approach involves generating a diverse and comprehensive list of 700 prompts for text-to-video generation.
Then, we evaluate the state-of-the-art video generative models on our carefully designed benchmark, in terms of visual qualities, content qualities, motion qualities, and text-video alignment with 17 well-selected objective metrics.
arXiv Detail & Related papers (2023-10-17T17:50:46Z) - Dual-Attention Model for Aspect-Level Sentiment Classification [0.0]
I propose a novel dual-attention model(DAM) for aspect-level sentiment classification.
We evaluate the proposed approach on three datasets: laptop and restaurant are from SemEval 2014, and the last one is a twitter dataset.
arXiv Detail & Related papers (2023-03-14T08:04:38Z) - Multitask Learning for Low Resource Spoken Language Understanding [26.106133114838215]
We train models on dual objectives with automatic speech recognition and intent classification or sentiment classification.
Our models, although being of modest size, show improvements over models trained end-to-end on intent classification.
We study the performance of the models in low-resource scenario by training the models with as few as one example per class.
arXiv Detail & Related papers (2022-11-24T16:38:17Z) - Incremental Online Learning Algorithms Comparison for Gesture and Visual
Smart Sensors [68.8204255655161]
This paper compares four state-of-the-art algorithms in two real applications: gesture recognition based on accelerometer data and image classification.
Our results confirm these systems' reliability and the feasibility of deploying them in tiny-memory MCUs.
arXiv Detail & Related papers (2022-09-01T17:05:20Z) - Revisiting Classifier: Transferring Vision-Language Models for Video
Recognition [102.93524173258487]
Transferring knowledge from task-agnostic pre-trained deep models for downstream tasks is an important topic in computer vision research.
In this study, we focus on transferring knowledge for video classification tasks.
We utilize the well-pretrained language model to generate good semantic target for efficient transferring learning.
arXiv Detail & Related papers (2022-07-04T10:00:47Z) - Long Document Summarization with Top-down and Bottom-up Inference [113.29319668246407]
We propose a principled inference framework to improve summarization models on two aspects.
Our framework assumes a hierarchical latent structure of a document where the top-level captures the long range dependency.
We demonstrate the effectiveness of the proposed framework on a diverse set of summarization datasets.
arXiv Detail & Related papers (2022-03-15T01:24:51Z) - Lightweight Object-level Topological Semantic Mapping and Long-term
Global Localization based on Graph Matching [19.706907816202946]
We present a novel lightweight object-level mapping and localization method with high accuracy and robustness.
We use object-level features with both semantic and geometric information to model landmarks in the environment.
Based on the proposed map, the robust localization is achieved by constructing a novel local semantic scene graph descriptor.
arXiv Detail & Related papers (2022-01-16T05:47:07Z) - LIDSNet: A Lightweight on-device Intent Detection model using Deep
Siamese Network [2.624902795082451]
LIDSNet is a novel lightweight on-device intent detection model.
We show that our model is at least 41x lighter and 30x faster during inference than MobileBERT on Samsung Galaxy S20 device.
arXiv Detail & Related papers (2021-10-06T18:20:37Z) - Enhancing the Generalization for Intent Classification and Out-of-Domain
Detection in SLU [70.44344060176952]
Intent classification is a major task in spoken language understanding (SLU)
Recent works have shown that using extra data and labels can improve the OOD detection performance.
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
arXiv Detail & Related papers (2021-06-28T08:27:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.