LMaaS: Exploring Pricing Strategy of Large Model as a Service for
Communication
- URL: http://arxiv.org/abs/2401.02675v1
- Date: Fri, 5 Jan 2024 07:19:19 GMT
- Title: LMaaS: Exploring Pricing Strategy of Large Model as a Service for
Communication
- Authors: Panlong Wu, Qi Liu, Yanjie Dong, Fangxin Wang
- Abstract summary: We argue that a pay-as-you-go service mode will be suitable in this context, referred to as Large Model as a Service (LM)
We propose an Iterative Model Pricing (IMP) algorithm that optimize the prices of large models iteratively by reasoning customers' future rental decisions.
In the second step, we optimize customers' selection decisions by designing a robust selecting and renting algorithm.
- Score: 11.337245234301857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The next generation of communication is envisioned to be intelligent
communication, that can replace traditional symbolic communication, where
highly condensed semantic information considering both source and channel will
be extracted and transmitted with high efficiency. The recent popular large
models such as GPT4 and the boosting learning techniques lay a solid foundation
for the intelligent communication, and prompt the practical deployment of it in
the near future. Given the characteristics of "training once and widely use" of
those multimodal large language models, we argue that a pay-as-you-go service
mode will be suitable in this context, referred to as Large Model as a Service
(LMaaS). However, the trading and pricing problem is quite complex with
heterogeneous and dynamic customer environments, making the pricing
optimization problem challenging in seeking on-hand solutions. In this paper,
we aim to fill this gap and formulate the LMaaS market trading as a Stackelberg
game with two steps. In the first step, we optimize the seller's pricing
decision and propose an Iterative Model Pricing (IMP) algorithm that optimizes
the prices of large models iteratively by reasoning customers' future rental
decisions, which is able to achieve a near-optimal pricing solution. In the
second step, we optimize customers' selection decisions by designing a robust
selecting and renting (RSR) algorithm, which is guaranteed to be optimal with
rigorous theoretical proof. Extensive experiments confirm the effectiveness and
robustness of our algorithms.
Related papers
- Optima: Optimizing Effectiveness and Efficiency for LLM-Based Multi-Agent System [75.25394449773052]
Large Language Model (LLM) based multi-agent systems (MAS) show remarkable potential in collaborative problem-solving.
Yet they still face critical challenges: low communication efficiency, poor scalability, and a lack of effective parameter-updating optimization methods.
We present Optima, a novel framework that addresses these issues by significantly enhancing both communication efficiency and task effectiveness.
arXiv Detail & Related papers (2024-10-10T17:00:06Z) - Cost-Effective Proxy Reward Model Construction with On-Policy and Active Learning [70.22819290458581]
Reinforcement learning with human feedback (RLHF) is a widely adopted approach in current large language model pipelines.
Our approach introduces two key innovations: (1) on-policy query to avoid OOD and imbalance issues in seed data, and (2) active learning to select the most informative data for preference queries.
arXiv Detail & Related papers (2024-07-02T10:09:19Z) - Neural Optimization with Adaptive Heuristics for Intelligent Marketing System [1.3079444139643954]
We propose a general framework for marketing AI systems, the Neural Optimization with Adaptive Heuristics (Noah) framework.
Noah is the first general framework for marketing optimization that considers both to-business (2B) and to-consumer (2C) products, as well as both owned and paid channels.
We describe key modules of the Noah framework, including prediction, optimization, and adaptive audiences, providing examples for bidding and content optimization.
arXiv Detail & Related papers (2024-05-17T01:44:30Z) - Optimising Calls to Large Language Models with Uncertainty-Based Two-Tier Selection [80.63946798650653]
Decision centers on whether to use a large LLM with better performance or a smaller one with reduced costs.
We propose a simpler solution; we use only the uncertainty of the generations of the small LLM as the decision criterion.
Our experiments reveal this simple solution optimally balances cost and performance, outperforming existing methods on 25 out of 27 experimental setups.
arXiv Detail & Related papers (2024-05-03T14:38:59Z) - Generative AI Agents with Large Language Model for Satellite Networks via a Mixture of Experts Transmission [74.10928850232717]
This paper develops generative artificial intelligence (AI) agents for model formulation and then applies a mixture of experts (MoE) to design transmission strategies.
Specifically, we leverage large language models (LLMs) to build an interactive modeling paradigm.
We propose an MoE-proximal policy optimization (PPO) approach to solve the formulated problem.
arXiv Detail & Related papers (2024-04-14T03:44:54Z) - From Large Language Models and Optimization to Decision Optimization
CoPilot: A Research Manifesto [2.4981381729038743]
We propose research at the intersection of Large Language Models and optimization to create a Decision Optimization CoPilot (DOCP)
DOCP is an AI tool designed to assist any decision maker, interacting in natural language to grasp the business problem, subsequently formulating and solving the corresponding optimization model.
We show that a) LLMs already provide substantial novel capabilities relevant to a DOCP, and b. major research challenges remain to be addressed.
arXiv Detail & Related papers (2024-02-26T03:10:11Z) - Large Language Models for Supply Chain Optimization [4.554094815136834]
We study how Large Language Models (LLMs) can help bridge the gap between supply chain automation and human comprehension and trust thereof.
We design OptiGuide -- a framework that accepts as input queries in plain text, and outputs insights about the underlying outcomes.
We demonstrate the effectiveness of our framework on a real server placement scenario within Microsoft's cloud supply chain.
arXiv Detail & Related papers (2023-07-08T01:42:22Z) - Sequential Information Design: Markov Persuasion Process and Its
Efficient Reinforcement Learning [156.5667417159582]
This paper proposes a novel model of sequential information design, namely the Markov persuasion processes (MPPs)
Planning in MPPs faces the unique challenge in finding a signaling policy that is simultaneously persuasive to the myopic receivers and inducing the optimal long-term cumulative utilities of the sender.
We design a provably efficient no-regret learning algorithm, the Optimism-Pessimism Principle for Persuasion Process (OP4), which features a novel combination of both optimism and pessimism principles.
arXiv Detail & Related papers (2022-02-22T05:41:43Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Multi-Purchase Behavior: Modeling, Estimation and Optimization [0.9337154228221861]
We present a parsimonious multi-purchase family of choice models called the Bundle-MVL-K family.
We develop a binary search based iterative strategy that efficiently computes optimized recommendations for this model.
arXiv Detail & Related papers (2020-06-14T23:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.