Modeling Product Search Relevance in e-Commerce
- URL: http://arxiv.org/abs/2001.04980v1
- Date: Tue, 14 Jan 2020 21:17:55 GMT
- Title: Modeling Product Search Relevance in e-Commerce
- Authors: Rahul Radhakrishnan Iyer, Rohan Kohli, Shrimai Prabhumoye
- Abstract summary: We propose a robust way of predicting relevance scores given a search query and a product.
We compare conventional information retrieval models such as BM25 and Indri with deep learning models such as word2vec, sentence2vec and paragraph2vec.
- Score: 7.139647051098728
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rapid growth of e-Commerce, online product search has emerged as a
popular and effective paradigm for customers to find desired products and
engage in online shopping. However, there is still a big gap between the
products that customers really desire to purchase and relevance of products
that are suggested in response to a query from the customer. In this paper, we
propose a robust way of predicting relevance scores given a search query and a
product, using techniques involving machine learning, natural language
processing and information retrieval. We compare conventional information
retrieval models such as BM25 and Indri with deep learning models such as
word2vec, sentence2vec and paragraph2vec. We share some of our insights and
findings from our experiments.
Related papers
- Language Models: A Guide for the Perplexed [51.88841610098437]
This tutorial aims to help narrow the gap between those who study language models and those who are intrigued and want to learn more.
We offer a scientific viewpoint that focuses on questions amenable to study through experimentation.
We situate language models as they are today in the context of the research that led to their development.
arXiv Detail & Related papers (2023-11-29T01:19:02Z) - Leveraging Large Language Models for Enhanced Product Descriptions in
eCommerce [6.318353155416729]
This paper introduces a novel methodology for automating product description generation using the LLAMA 2.0 7B language model.
We train the model on a dataset of authentic product descriptions from Walmart, one of the largest eCommerce platforms.
Our findings reveal that the system is not only scalable but also significantly reduces the human workload involved in creating product descriptions.
arXiv Detail & Related papers (2023-10-24T00:55:14Z) - Product Information Extraction using ChatGPT [69.12244027050454]
This paper explores the potential of ChatGPT for extracting attribute/value pairs from product descriptions.
Our results show that ChatGPT achieves a performance similar to a pre-trained language model but requires much smaller amounts of training data and computation for fine-tuning.
arXiv Detail & Related papers (2023-06-23T09:30:01Z) - Automated Extraction of Fine-Grained Standardized Product Information
from Unstructured Multilingual Web Data [66.21317300595483]
We show how recent advances in machine learning, combined with a recently published multilingual data set, enable robust product attribute extraction.
Our models can reliably predict product attributes across online shops, languages, or both.
arXiv Detail & Related papers (2023-02-23T16:26:11Z) - Intent-based Product Collections for E-commerce using Pretrained
Language Models [8.847005669899703]
We use a pretrained language model (PLM) that leverages textual attributes of web-scale products to make intent-based product collections.
Our model significantly outperforms the search-based baseline model for intent-based product matching in offline evaluations.
Online experimental results on our e-commerce platform show that the PLM-based method can construct collections of products with increased CTR, CVR, and order-diversity compared to expert-crafted collections.
arXiv Detail & Related papers (2021-10-15T17:52:42Z) - Aggregated Customer Engagement Model [0.571097144710995]
E-commerce websites use machine learned ranking models to serve shopping results to customers.
New or under-impressed products do not have enough customer engagement signals and end up at a disadvantage when being ranked alongside popular products.
We propose a novel method for data curation that aggregates all customer engagements within a day for the same query to use as input training data.
arXiv Detail & Related papers (2021-08-17T20:58:10Z) - Query2Prod2Vec Grounded Word Embeddings for eCommerce [4.137464623395377]
We present a model that grounds lexical representations for product search in product embeddings.
We leverage shopping sessions to learn the underlying space and use merchandising annotations to build lexical analogies for evaluation.
arXiv Detail & Related papers (2021-04-02T21:32:43Z) - Dual Side Deep Context-aware Modulation for Social Recommendation [50.59008227281762]
We propose a novel graph neural network to model the social relation and collaborative relation.
On top of high-order relations, a dual side deep context-aware modulation is introduced to capture the friends' information and item attraction.
arXiv Detail & Related papers (2021-03-16T11:08:30Z) - User-Inspired Posterior Network for Recommendation Reason Generation [53.035224183349385]
Recommendation reason generation plays a vital role in attracting customers' attention as well as improving user experience.
We propose a user-inspired multi-source posterior transformer (MSPT), which induces the model reflecting the users' interests.
Experimental results show that our model is superior to traditional generative models.
arXiv Detail & Related papers (2021-02-16T02:08:52Z) - Active Learning for Product Type Ontology Enhancement in E-commerce [16.170442845801183]
We propose an active learning framework that efficiently utilizes domain experts' knowledge for PT discovery.
We also show the quality and coverage of the resulting PTs in the experiment results.
arXiv Detail & Related papers (2020-09-19T02:21:12Z) - E-BERT: A Phrase and Product Knowledge Enhanced Language Model for
E-commerce [63.333860695727424]
E-commerce tasks require accurate understanding of domain phrases, whereas such fine-grained phrase-level knowledge is not explicitly modeled by BERT's training objective.
To tackle the problem, we propose a unified pre-training framework, namely, E-BERT.
Specifically, to preserve phrase-level knowledge, we introduce Adaptive Hybrid Masking, which allows the model to adaptively switch from learning preliminary word knowledge to learning complex phrases.
To utilize product-level knowledge, we introduce Neighbor Product Reconstruction, which trains E-BERT to predict a product's associated neighbors with a denoising cross attention layer
arXiv Detail & Related papers (2020-09-07T00:15:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.