An Interpretable Ensemble of Graph and Language Models for Improving
Search Relevance in E-Commerce
- URL: http://arxiv.org/abs/2403.00923v1
- Date: Fri, 1 Mar 2024 19:08:25 GMT
- Title: An Interpretable Ensemble of Graph and Language Models for Improving
Search Relevance in E-Commerce
- Authors: Nurendra Choudhary, Edward W Huang, Karthik Subbian, Chandan K. Reddy
- Abstract summary: We propose Plug and Play Graph LAnguage Model (PP-GLAM), an explainable ensemble of plug and play models.
Our approach uses a modular framework with uniform data processing pipelines.
We show that PP-GLAM outperforms several state-of-the-art baselines and a proprietary model on real-world multilingual, multi-regional e-commerce datasets.
- Score: 22.449320058423886
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The problem of search relevance in the E-commerce domain is a challenging one
since it involves understanding the intent of a user's short nuanced query and
matching it with the appropriate products in the catalog. This problem has
traditionally been addressed using language models (LMs) and graph neural
networks (GNNs) to capture semantic and inter-product behavior signals,
respectively. However, the rapid development of new architectures has created a
gap between research and the practical adoption of these techniques. Evaluating
the generalizability of these models for deployment requires extensive
experimentation on complex, real-world datasets, which can be non-trivial and
expensive. Furthermore, such models often operate on latent space
representations that are incomprehensible to humans, making it difficult to
evaluate and compare the effectiveness of different models. This lack of
interpretability hinders the development and adoption of new techniques in the
field. To bridge this gap, we propose Plug and Play Graph LAnguage Model
(PP-GLAM), an explainable ensemble of plug and play models. Our approach uses a
modular framework with uniform data processing pipelines. It employs additive
explanation metrics to independently decide whether to include (i) language
model candidates, (ii) GNN model candidates, and (iii) inter-product behavioral
signals. For the task of search relevance, we show that PP-GLAM outperforms
several state-of-the-art baselines as well as a proprietary model on real-world
multilingual, multi-regional e-commerce datasets. To promote better model
comprehensibility and adoption, we also provide an analysis of the
explainability and computational complexity of our model. We also provide the
public codebase and provide a deployment strategy for practical implementation.
Related papers
- Matchmaker: Self-Improving Large Language Model Programs for Schema Matching [60.23571456538149]
We propose a compositional language model program for schema matching, comprised of candidate generation, refinement and confidence scoring.
Matchmaker self-improves in a zero-shot manner without the need for labeled demonstrations.
Empirically, we demonstrate on real-world medical schema matching benchmarks that Matchmaker outperforms previous ML-based approaches.
arXiv Detail & Related papers (2024-10-31T16:34:03Z) - Improving the Capabilities of Large Language Model Based Marketing Analytics Copilots With Semantic Search And Fine-Tuning [0.9787137564521711]
We show how a combination of semantic search, prompt engineering, and fine-tuning can be applied to dramatically improve the ability of LLMs to execute these tasks accurately.
We compare both proprietary models, like GPT-4, and open-source models, like Llama-2-70b, as well as various embedding methods.
arXiv Detail & Related papers (2024-04-16T03:39:16Z) - GLaM: Fine-Tuning Large Language Models for Domain Knowledge Graph Alignment via Neighborhood Partitioning and Generative Subgraph Encoding [39.67113788660731]
We introduce a framework for developing Graph-aligned LAnguage Models (GLaM)
We demonstrate that grounding the models in specific graph-based knowledge expands the models' capacity for structure-based reasoning.
arXiv Detail & Related papers (2024-02-09T19:53:29Z) - Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning [79.53130089003986]
Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
arXiv Detail & Related papers (2023-10-05T09:09:44Z) - StableLLaVA: Enhanced Visual Instruction Tuning with Synthesized
Image-Dialogue Data [129.92449761766025]
We propose a novel data collection methodology that synchronously synthesizes images and dialogues for visual instruction tuning.
This approach harnesses the power of generative models, marrying the abilities of ChatGPT and text-to-image generative models.
Our research includes comprehensive experiments conducted on various datasets.
arXiv Detail & Related papers (2023-08-20T12:43:52Z) - Enhancing Knowledge Graph Construction Using Large Language Models [0.0]
This paper analyzes how the current advances in foundational LLM, like ChatGPT, can be compared with the specialized pretrained models, like REBEL, for joint entity and relation extraction.
We created pipelines for the automatic creation of Knowledge Graphs from raw texts, and our findings indicate that using advanced LLM models can improve the accuracy of the process of creating these graphs from unstructured text.
arXiv Detail & Related papers (2023-05-08T12:53:06Z) - Scaling Vision-Language Models with Sparse Mixture of Experts [128.0882767889029]
We show that mixture-of-experts (MoE) techniques can achieve state-of-the-art performance on a range of benchmarks over dense models of equivalent computational cost.
Our research offers valuable insights into stabilizing the training of MoE models, understanding the impact of MoE on model interpretability, and balancing the trade-offs between compute performance when scaling vision-language models.
arXiv Detail & Related papers (2023-03-13T16:00:31Z) - DIME: Fine-grained Interpretations of Multimodal Models via Disentangled
Local Explanations [119.1953397679783]
We focus on advancing the state-of-the-art in interpreting multimodal models.
Our proposed approach, DIME, enables accurate and fine-grained analysis of multimodal models.
arXiv Detail & Related papers (2022-03-03T20:52:47Z) - ARM-Net: Adaptive Relation Modeling Network for Structured Data [29.94433633729326]
ARM-Net is an adaptive relation modeling network tailored for structured data and a lightweight framework ARMOR based on ARM-Net for relational data.
We show that ARM-Net consistently outperforms existing models and provides more interpretable predictions for datasets.
arXiv Detail & Related papers (2021-07-05T07:37:24Z) - Edge-assisted Democratized Learning Towards Federated Analytics [67.44078999945722]
We show the hierarchical learning structure of the proposed edge-assisted democratized learning mechanism, namely Edge-DemLearn.
We also validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions.
arXiv Detail & Related papers (2020-12-01T11:46:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.