Handling Long-Tail Queries with Slice-Aware Conversational Systems
- URL: http://arxiv.org/abs/2104.13216v1
- Date: Mon, 26 Apr 2021 10:23:28 GMT
- Title: Handling Long-Tail Queries with Slice-Aware Conversational Systems
- Authors: Cheng Wang, Sun Kim, Taiwoo Park, Sajal Choudhary, Sunghyun Park,
Young-Bum Kim, Ruhi Sarikaya, Sungjin Lee
- Abstract summary: We explore the recent concept of slice-based learning (SBL)
We first define a set of labeling functions to generate weak supervision data for the tail intents.
We then extend the baseline model towards a slice, which monitors and improves the model performance on the selected tail intents.
Experiments show that the slice-aware model is beneficial in improving model performance for the tail intents while maintaining the overall performance.
- Score: 29.41747850530487
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We have been witnessing the usefulness of conversational AI systems such as
Siri and Alexa, directly impacting our daily lives. These systems normally rely
on machine learning models evolving over time to provide quality user
experience. However, the development and improvement of the models are
challenging because they need to support both high (head) and low (tail) usage
scenarios, requiring fine-grained modeling strategies for specific data subsets
or slices. In this paper, we explore the recent concept of slice-based learning
(SBL) (Chen et al., 2019) to improve our baseline conversational skill routing
system on the tail yet critical query traffic. We first define a set of
labeling functions to generate weak supervision data for the tail intents. We
then extend the baseline model towards a slice-aware architecture, which
monitors and improves the model performance on the selected tail intents.
Applied to de-identified live traffic from a commercial conversational AI
system, our experiments show that the slice-aware model is beneficial in
improving model performance for the tail intents while maintaining the overall
performance.
Related papers
- A Two-Phase Recall-and-Select Framework for Fast Model Selection [13.385915962994806]
We propose a two-phase (coarse-recall and fine-selection) model selection framework.
It aims to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets.
It has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods.
arXiv Detail & Related papers (2024-03-28T14:44:44Z) - AIDE: An Automatic Data Engine for Object Detection in Autonomous Driving [68.73885845181242]
We propose an Automatic Data Engine (AIDE) that automatically identifies issues, efficiently curates data, improves the model through auto-labeling, and verifies the model through generation of diverse scenarios.
We further establish a benchmark for open-world detection on AV datasets to comprehensively evaluate various learning paradigms, demonstrating our method's superior performance at a reduced cost.
arXiv Detail & Related papers (2024-03-26T04:27:56Z) - Data-efficient Large Vision Models through Sequential Autoregression [58.26179273091461]
We develop an efficient, autoregression-based vision model on a limited dataset.
We demonstrate how this model achieves proficiency in a spectrum of visual tasks spanning both high-level and low-level semantic understanding.
Our empirical evaluations underscore the model's agility in adapting to various tasks, heralding a significant reduction in the parameter footprint.
arXiv Detail & Related papers (2024-02-07T13:41:53Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Retrieval-Enhanced Contrastive Vision-Text Models [61.783728119255365]
We propose to equip vision-text models with the ability to refine their embedding with cross-modal retrieved information from a memory at inference time.
Remarkably, we show that this can be done with a light-weight, single-layer, fusion transformer on top of a frozen CLIP.
Our experiments validate that our retrieval-enhanced contrastive (RECO) training improves CLIP performance substantially on several challenging fine-grained tasks.
arXiv Detail & Related papers (2023-06-12T15:52:02Z) - Knowledge Transfer For On-Device Speech Emotion Recognition with Neural
Structured Learning [19.220263739291685]
Speech emotion recognition (SER) has been a popular research topic in human-computer interaction (HCI)
We propose a neural structured learning (NSL) framework through building synthesized graphs.
Our experiments demonstrate that training a lightweight SER model on the target dataset with speech samples and graphs can not only produce small SER models, but also enhance the model performance.
arXiv Detail & Related papers (2022-10-26T18:38:42Z) - Amazon SageMaker Model Monitor: A System for Real-Time Insights into
Deployed Machine Learning Models [15.013638492229376]
We present Amazon SageMaker Model Monitor, a fully managed service that continuously monitors the quality of machine learning models hosted on Amazon SageMaker.
Our system automatically detects data, concept, bias, and feature attribution drift in models in real-time and provides alerts so that model owners can take corrective actions.
arXiv Detail & Related papers (2021-11-26T18:35:38Z) - STAR: Sparse Transformer-based Action Recognition [61.490243467748314]
This work proposes a novel skeleton-based human action recognition model with sparse attention on the spatial dimension and segmented linear attention on the temporal dimension of data.
Experiments show that our model can achieve comparable performance while utilizing much less trainable parameters and achieve high speed in training and inference.
arXiv Detail & Related papers (2021-07-15T02:53:11Z) - Learning Transferrable Parameters for Long-tailed Sequential User
Behavior Modeling [70.64257515361972]
We argue that focusing on tail users could bring more benefits and address the long tails issue.
Specifically, we propose a gradient alignment and adopt an adversarial training scheme to facilitate knowledge transfer from the head to the tail.
arXiv Detail & Related papers (2020-10-22T03:12:02Z) - Self-supervised Learning for Large-scale Item Recommendations [18.19202958502061]
Large scale recommender models find most relevant items from huge catalogs.
With millions to billions of items in the corpus, users tend to provide feedback for a very small set of them.
We propose a multi-task self-supervised learning framework for large-scale item recommendations.
arXiv Detail & Related papers (2020-07-25T06:21:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.