Visually Similar Products Retrieval for Shopsy
- URL: http://arxiv.org/abs/2210.04560v1
- Date: Mon, 10 Oct 2022 10:59:18 GMT
- Title: Visually Similar Products Retrieval for Shopsy
- Authors: Prajit Nadkarni, Narendra Varma Dasararaju
- Abstract summary: We design a visual search system for reseller commerce using a multi-task learning approach.
Our model consists of three different tasks: attribute classification, triplet ranking and variational autoencoder (VAE)
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual search is of great assistance in reseller commerce, especially for
non-tech savvy users with affinity towards regional languages. It allows
resellers to accurately locate the products that they seek, unlike textual
search which recommends products from head brands. Product attributes available
in e-commerce have a great potential for building better visual search systems
as they capture fine grained relations between data points. In this work, we
design a visual search system for reseller commerce using a multi-task learning
approach. We also highlight and address the challenges like image compression,
cropping, scribbling on the image, etc, faced in reseller commerce. Our model
consists of three different tasks: attribute classification, triplet ranking
and variational autoencoder (VAE). Masking technique is used for designing the
attribute classification. Next, we introduce an offline triplet mining
technique which utilizes information from multiple attributes to capture
relative order within the data. This technique displays a better performance
compared to the traditional triplet mining baseline, which uses single
label/attribute information. We also compare and report incremental gain
achieved by our unified multi-task model over each individual task separately.
The effectiveness of our method is demonstrated using the in-house dataset of
product images from the Lifestyle business-unit of Flipkart, India's largest
e-commerce company. To efficiently retrieve the images in production, we use
the Approximate Nearest Neighbor (ANN) index. Finally, we highlight our
production environment constraints and present the design choices and
experiments conducted to select a suitable ANN index.
Related papers
- PAE: LLM-based Product Attribute Extraction for E-Commerce Fashion Trends [0.6445605125467574]
This paper presents PAE, a product attribute extraction algorithm for future trend reports consisting text and images in PDF format.
Our contributions are three-fold: (a) We develop PAE, an efficient framework to extract attributes from unstructured data (text and images); (b) We provide catalog matching methodology based on BERT representations to discover the existing attributes using upcoming attribute values; (c) We conduct extensive experiments with several baselines and show that PAE is an effective, flexible and on par or superior (avg 92.5% F1-Score) framework to existing state-of-the-art for attribute value extraction
arXiv Detail & Related papers (2024-05-27T17:50:25Z) - Enhanced E-Commerce Attribute Extraction: Innovating with Decorative
Relation Correction and LLAMA 2.0-Based Annotation [4.81846973621209]
We propose a pioneering framework that integrates BERT for classification, a Conditional Random Fields (CRFs) layer for attribute value extraction, and Large Language Models (LLMs) for data annotation.
Our approach capitalizes on the robust representation learning of BERT, synergized with the sequence decoding prowess of CRFs, to adeptly identify and extract attribute values.
Our methodology is rigorously validated on various datasets, including Walmart, BestBuy's e-commerce NER dataset, and the CoNLL dataset.
arXiv Detail & Related papers (2023-12-09T08:26:30Z) - Product Information Extraction using ChatGPT [69.12244027050454]
This paper explores the potential of ChatGPT for extracting attribute/value pairs from product descriptions.
Our results show that ChatGPT achieves a performance similar to a pre-trained language model but requires much smaller amounts of training data and computation for fine-tuning.
arXiv Detail & Related papers (2023-06-23T09:30:01Z) - DATE: Domain Adaptive Product Seeker for E-commerce [75.25578276795383]
Product Retrieval (PR) and Grounding (PG) aim to seek image and object-level products respectively according to a textual query.
We propose a bf Domain bf Adaptive Producbf t Sbf eeker (bf DATE) framework, regarding PR and PG as Product Seeking problem at different levels.
arXiv Detail & Related papers (2023-04-07T14:40:16Z) - e-CLIP: Large-Scale Vision-Language Representation Learning in
E-commerce [9.46186546774799]
We propose a contrastive learning framework that aligns language and visual models using unlabeled raw product text and images.
We present techniques we used to train large-scale representation learning models and share solutions that address domain-specific challenges.
arXiv Detail & Related papers (2022-07-01T05:16:47Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - ItemSage: Learning Product Embeddings for Shopping Recommendations at
Pinterest [60.841761065439414]
At Pinterest, we build a single set of product embeddings called ItemSage to provide relevant recommendations in all shopping use cases.
This approach has led to significant improvements in engagement and conversion metrics, while reducing both infrastructure and maintenance cost.
arXiv Detail & Related papers (2022-05-24T02:28:58Z) - ACE-BERT: Adversarial Cross-modal Enhanced BERT for E-commerce Retrieval [6.274310862007448]
We propose a novel Adrial Cross-modal Enhanced BERT (ACE-BERT) for efficient E-commerce retrieval.
With the pre-trained enhanced BERT as the backbone network, ACE-BERT adopts adversarial learning to ensure the distribution consistency of different modality representations.
Experimental results demonstrate that ACE-BERT outperforms the state-of-the-art approaches on the retrieval task.
arXiv Detail & Related papers (2021-12-14T07:36:20Z) - Pre-training Graph Transformer with Multimodal Side Information for
Recommendation [82.4194024706817]
We propose a pre-training strategy to learn item representations by considering both item side information and their relationships.
We develop a novel sampling algorithm named MCNSampling to select contextual neighbors for each item.
The proposed Pre-trained Multimodal Graph Transformer (PMGT) learns item representations with two objectives: 1) graph structure reconstruction, and 2) masked node feature reconstruction.
arXiv Detail & Related papers (2020-10-23T10:30:24Z) - Automatic Validation of Textual Attribute Values in E-commerce Catalog
by Learning with Limited Labeled Data [61.789797281676606]
We propose a novel meta-learning latent variable approach, called MetaBridge.
It can learn transferable knowledge from a subset of categories with limited labeled data.
It can capture the uncertainty of never-seen categories with unlabeled data.
arXiv Detail & Related papers (2020-06-15T21:31:05Z) - Large-scale Real-time Personalized Similar Product Recommendations [28.718371564543517]
We introduce our real-time personalized algorithm to model product similarity and real-time user interests.
Our method achieves a 10% improvement on the add-cart number in the real-world e-commerce scenario.
arXiv Detail & Related papers (2020-04-12T23:16:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.