Distantly Supervised Transformers For E-Commerce Product QA
- URL: http://arxiv.org/abs/2104.02947v1
- Date: Wed, 7 Apr 2021 06:37:16 GMT
- Title: Distantly Supervised Transformers For E-Commerce Product QA
- Authors: Happy Mittal, Aniket Chakrabarti, Belhassen Bayar, Animesh Anant
Sharma, Nikhil Rasiwasia
- Abstract summary: We propose a practical instant question answering (QA) system on product pages of ecommerce services.
For each user query, relevant community question answer (CQA) pairs are retrieved.
Our proposed transformer-based model learns a robust relevance function by jointly learning unified syntactic and semantic representations.
- Score: 5.460297795256275
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a practical instant question answering (QA) system on product
pages of ecommerce services, where for each user query, relevant community
question answer (CQA) pairs are retrieved. User queries and CQA pairs differ
significantly in language characteristics making relevance learning difficult.
Our proposed transformer-based model learns a robust relevance function by
jointly learning unified syntactic and semantic representations without the
need for human labeled data. This is achieved by distantly supervising our
model by distilling from predictions of a syntactic matching system on user
queries and simultaneously training with CQA pairs. Training with CQA pairs
helps our model learning semantic QA relevance and distant supervision enables
learning of syntactic features as well as the nuances of user querying
language. Additionally, our model encodes queries and candidate responses
independently allowing offline candidate embedding generation thereby
minimizing the need for real-time transformer model execution. Consequently,
our framework is able to scale to large e-commerce QA traffic. Extensive
evaluation on user queries shows that our framework significantly outperforms
both syntactic and semantic baselines in offline as well as large scale online
A/B setups of a popular e-commerce service.
Related papers
- Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation [65.16137964758612]
We explore the use of long-context capabilities in large language models to create synthetic reading comprehension data from entire books.
Our objective is to test the capabilities of LLMs to analyze, understand, and reason over problems that require a detailed comprehension of long spans of text.
arXiv Detail & Related papers (2024-05-31T20:15:10Z) - Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity [59.57065228857247]
Retrieval-augmented Large Language Models (LLMs) have emerged as a promising approach to enhancing response accuracy in several tasks, such as Question-Answering (QA)
We propose a novel adaptive QA framework, that can dynamically select the most suitable strategy for (retrieval-augmented) LLMs based on the query complexity.
We validate our model on a set of open-domain QA datasets, covering multiple query complexities, and show that ours enhances the overall efficiency and accuracy of QA systems.
arXiv Detail & Related papers (2024-03-21T13:52:30Z) - Improving Text Matching in E-Commerce Search with A Rationalizable,
Intervenable and Fast Entity-Based Relevance Model [78.80174696043021]
We propose a novel model called the Entity-Based Relevance Model (EBRM)
The decomposition allows us to use a Cross-encoder QE relevance module for high accuracy.
We also show that pretraining the QE module with auto-generated QE data from user logs can further improve the overall performance.
arXiv Detail & Related papers (2023-07-01T15:44:53Z) - PENTATRON: PErsonalized coNText-Aware Transformer for Retrieval-based
cOnversational uNderstanding [18.788620612619823]
In a large fraction of the global traffic from customers using smart digital assistants, frictions in dialogues may be attributed to incorrect understanding.
We build and evaluate a scalable entity correction system, PENTATRON.
We show a significant upward movement of the key metric (Exact Match) by up to 500.97%.
arXiv Detail & Related papers (2022-10-22T00:14:47Z) - Conversational Query Rewriting with Self-supervised Learning [36.392717968127016]
Conversational Query Rewriting (CQR) aims to simplify the multi-turn dialogue modeling into a single-turn problem by explicitly rewriting the conversational query into a self-contained utterance.
Existing approaches rely on massive supervised training data, which is labor-intensive to annotate.
We propose to construct a large-scale CQR dataset automatically via self-supervised learning, which does not need human annotation.
arXiv Detail & Related papers (2021-02-09T08:57:53Z) - IQA: Interactive Query Construction in Semantic Question Answering
Systems [8.961129460639999]
We introduce IQA - an interaction scheme for SQA pipelines.
We show that even a small number of user interactions can lead to significant improvements in the performance of SQA systems.
arXiv Detail & Related papers (2020-06-20T10:02:20Z) - Mining Implicit Relevance Feedback from User Behavior for Web Question
Answering [92.45607094299181]
We make the first study to explore the correlation between user behavior and passage relevance.
Our approach significantly improves the accuracy of passage ranking without extra human labeled data.
In practice, this work has proved effective to substantially reduce the human labeling cost for the QA service in a global commercial search engine.
arXiv Detail & Related papers (2020-06-13T07:02:08Z) - Generating Diverse and Consistent QA pairs from Contexts with
Information-Maximizing Hierarchical Conditional VAEs [62.71505254770827]
We propose a conditional variational autoencoder (HCVAE) for generating QA pairs given unstructured texts as contexts.
Our model obtains impressive performance gains over all baselines on both tasks, using only a fraction of data for training.
arXiv Detail & Related papers (2020-05-28T08:26:06Z) - Template-Based Question Generation from Retrieved Sentences for Improved
Unsupervised Question Answering [98.48363619128108]
We propose an unsupervised approach to training QA models with generated pseudo-training data.
We show that generating questions for QA training by applying a simple template on a related, retrieved sentence rather than the original context sentence improves downstream QA performance.
arXiv Detail & Related papers (2020-04-24T17:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.