EC-Guide: A Comprehensive E-Commerce Guide for Instruction Tuning and Quantization
- URL: http://arxiv.org/abs/2408.02970v1
- Date: Tue, 6 Aug 2024 05:50:41 GMT
- Title: EC-Guide: A Comprehensive E-Commerce Guide for Instruction Tuning and Quantization
- Authors: Zhaopeng Feng, Zijie Meng, Zuozhu Liu,
- Abstract summary: EC-Guide hrefhttps://github.com/fzp0424/EC-Guide-KDDUP-2024 is a comprehensive e-commerce guide for instruction tuning and quantization of LLMs.
Our solution is model-agnostic, enabling effective scalability across larger systems.
- Score: 7.982538359035973
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large language models (LLMs) have attracted considerable attention in various fields for their cost-effective solutions to diverse challenges, especially with advancements in instruction tuning and quantization. E-commerce, with its complex tasks and extensive product-user interactions, presents a promising application area for LLMs. However, the domain-specific concepts and knowledge inherent in e-commerce pose significant challenges for adapting general LLMs. To address this issue, we developed EC-Guide \href{https://github.com/fzp0424/EC-Guide-KDDUP-2024}, a comprehensive e-commerce guide for instruction tuning and quantization of LLMs. We also heuristically integrated Chain-of-Thought (CoT) during inference to enhance arithmetic performance. Our approach achieved the 2nd place in Track 2 and 5th place in Track 5 at the Amazon KDD Cup'24 \href{https://www.aicrowd.com/challenges/amazon-kdd-cup-2024-multi-task-online-shopping-challenge-for-llms }. Additionally, our solution is model-agnostic, enabling effective scalability across larger systems.
Related papers
- SEQ+MD: Learning Multi-Task as a SEQuence with Multi-Distribution Data [5.069855142454979]
We propose the SEQ+MD framework, which integrates sequential learning for multi-task learning (MTL) and feature-generated region-mask for multi-distribution input.
We show a strong increase on the high-value engagement including add-to-cart and purchase while keeping click performance neutral.
Our multi-regional learning module is "plug-and-play" and can be easily adapted to enhance other MTL applications.
arXiv Detail & Related papers (2024-08-23T20:14:27Z) - LLaSA: Large Language and E-Commerce Shopping Assistant [17.53318263751155]
We create an instruction dataset comprising 65,000 samples and diverse tasks, termed as EshopInstruct.
Through instruction tuning on our dataset, the assistant, named LLaSA, demonstrates the potential to function as an omnipotent assistant.
In the Amazon KDD Cup 2024 Challenge, our proposed method, LLaSA, achieved an overall ranking of 3rd place on ShopBench.
arXiv Detail & Related papers (2024-08-04T12:10:51Z) - Fine-tuning Multimodal Large Language Models for Product Bundling [53.01642741096356]
We introduce Bundle-MLLM, a novel framework that fine-tunes large language models (LLMs) through a hybrid item tokenization approach.
Specifically, we integrate textual, media, and relational data into a unified tokenization, introducing a soft separation token to distinguish between textual and non-textual tokens.
We propose a progressive optimization strategy that fine-tunes LLMs for disentangled objectives: 1) learning bundle patterns and 2) enhancing multimodal semantic understanding specific to product bundling.
arXiv Detail & Related papers (2024-07-16T13:30:14Z) - SoupLM: Model Integration in Large Language and Multi-Modal Models [51.12227693121004]
Training large language models (LLMs) requires significant computing resources.
Existing publicly available LLMs are typically pre-trained on diverse, privately curated datasets spanning various tasks.
arXiv Detail & Related papers (2024-07-11T05:38:15Z) - JungleGPT: Designing and Optimizing Compound AI Systems for E-Commerce [0.4194758891177309]
We present JungleGPT, the first compound AI system tailored for real-world e-commerce applications.
We outline the system's design and the techniques used to optimize its performance for practical use cases.
arXiv Detail & Related papers (2024-05-28T20:14:18Z) - TAT-LLM: A Specialized Language Model for Discrete Reasoning over Tabular and Textual Data [73.29220562541204]
We consider harnessing the amazing power of language models (LLMs) to solve our task.
We develop a TAT-LLM language model by fine-tuning LLaMA 2 with the training data generated automatically from existing expert-annotated datasets.
arXiv Detail & Related papers (2024-01-24T04:28:50Z) - EcomGPT-CT: Continual Pre-training of E-commerce Large Language Models
with Semi-structured Data [67.8302955948861]
Large Language Models (LLMs) pre-trained on massive corpora have exhibited remarkable performance on various NLP tasks.
Applying these models to specific domains still poses significant challenges, such as lack of domain knowledge.
We focus on domain-specific continual pre-training of LLMs using E-commerce domain as an exemplar.
arXiv Detail & Related papers (2023-12-25T11:31:47Z) - EcomGPT: Instruction-tuning Large Language Models with Chain-of-Task
Tasks for E-commerce [68.72104414369635]
We propose the first e-commerce instruction dataset EcomInstruct, with a total of 2.5 million instruction data.
EcomGPT outperforms ChatGPT in term of cross-dataset/task generalization on E-commerce tasks.
arXiv Detail & Related papers (2023-08-14T06:49:53Z) - How Can Recommender Systems Benefit from Large Language Models: A Survey [82.06729592294322]
Large language models (LLM) have shown impressive general intelligence and human-like capabilities.
We conduct a comprehensive survey on this research direction from the perspective of the whole pipeline in real-world recommender systems.
arXiv Detail & Related papers (2023-06-09T11:31:50Z) - Learning to Minimize Cost-to-Serve for Multi-Node Multi-Product Order
Fulfilment in Electronic Commerce [3.3865605512957457]
We find that the cost of delivery of products from the most node in the supply chain is a key challenge.
The large scale, highproblemity, and large geographical spread of e-commerce supply chains make this setting ideal for a carefully designed data-driven decision-making algorithm.
We show that a reinforcement learning based algorithm is competitive with these policies, with the potential of efficient scale-up in the real world.
arXiv Detail & Related papers (2021-12-16T09:42:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.