MERIT Feedback Elicits Better Bargaining in LLM Negotiators
- URL: http://arxiv.org/abs/2602.10467v2
- Date: Thu, 12 Feb 2026 01:41:02 GMT
- Title: MERIT Feedback Elicits Better Bargaining in LLM Negotiators
- Authors: Jihwan Oh, Murad Aghazada, Yooju Shin, Se-Young Yun, Taehyeon Kim,
- Abstract summary: AgoraBench is a new benchmark spanning nine challenging settings.<n>This is operationalized via agent utility, negotiation power, and acquisition ratio that implicitly measure how well the negotiation aligns with human preference.<n>Our mechanism substantially improves negotiation performance, yielding deeper strategic behavior and stronger opponent awareness.
- Score: 38.1466669265123
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bargaining is often regarded as a logical arena rather than an art or a matter of intuition, yet Large Language Models (LLMs) still struggle to navigate it due to limited strategic depth and difficulty adapting to complex human factors. Current benchmarks rarely capture this limitation. To bridge this gap, we present an utility feedback centric framework. Our contributions are: (i) AgoraBench, a new benchmark spanning nine challenging settings (e.g., deception, monopoly) that supports diverse strategy modeling; (ii) human-aligned, economically grounded metrics derived from utility theory. This is operationalized via agent utility, negotiation power, and acquisition ratio that implicitly measure how well the negotiation aligns with human preference and (iii) a human preference grounded dataset with learning pipeline that strengthens LLMs' bargaining ability through both prompting and finetuning. Empirical results indicate that baseline LLM strategies often diverge from human preferences, while our mechanism substantially improves negotiation performance, yielding deeper strategic behavior and stronger opponent awareness.
Related papers
- LLM Rationalis? Measuring Bargaining Capabilities of AI Negotiators [2.1952520391635586]
Bilateral negotiation is a complex, context-sensitive task in which human negotiators dynamically adjust anchors, pacing, and flexibility to exploit power asymmetries and informal cues.<n>We introduce a unified mathematical framework for modeling concession dynamics based on a hyperbolic tangent curve.<n>We conduct a large-scale empirical comparison between human negotiators and four state-of-the-art large language models (LLMs) across natural-language and numeric-offers settings.
arXiv Detail & Related papers (2025-12-15T07:50:09Z) - How Far Can LLMs Emulate Human Behavior?: A Strategic Analysis via the Buy-and-Sell Negotiation Game [0.8353024005684598]
This work proposes a methodology to quantitatively evaluate the human emotional and behavioral imitation and strategic decision-making capabilities of Large Language Models (LLMs)<n>Specifically, we assign different personas to multiple LLMs and conduct negotiations between a Buyer and a Seller, comprehensively analyzing outcomes such as win rates, transaction prices, and SHAP values.<n>Our experimental results show that models with higher existing benchmark scores tend to achieve better negotiation performance overall.
arXiv Detail & Related papers (2025-11-22T09:07:29Z) - EQ-Negotiator: Dynamic Emotional Personas Empower Small Language Models for Edge-Deployable Credit Negotiation [66.09161596959771]
Small language models (SLMs) offer a practical alternative, but suffer from a significant performance gap compared to large language models (LLMs)<n>This paper introduces EQ-Negotiator, a novel framework that bridges this capability gap using emotional personas.<n>We show that a 7B parameter language model with EQ-Negotiator achieves better debt recovery and negotiation efficiency than baseline LLMs more than 10 times its size.
arXiv Detail & Related papers (2025-11-05T11:25:07Z) - EvoEmo: Towards Evolved Emotional Policies for Adversarial LLM Agents in Multi-Turn Price Negotiation [61.627248012799704]
Existing Large Language Models (LLMs) agents largely overlook the functional role of emotions in such negotiations.<n>We present EvoEmo, an evolutionary reinforcement learning framework that optimize dynamic emotional expression in negotiations.
arXiv Detail & Related papers (2025-09-04T15:23:58Z) - LLM Agents for Bargaining with Utility-based Feedback [23.357706450282002]
We introduce a comprehensive framework centered on utility-based feedback.<n>Our contributions are threefold: (1) BargainArena, a novel benchmark dataset; (2) human-aligned, economically-grounded evaluation metrics inspired by utility theory; and (3) a structured feedback mechanism enabling LLMs to iteratively refine their bargaining strategies.
arXiv Detail & Related papers (2025-05-29T02:07:27Z) - EmoDebt: Bayesian-Optimized Emotional Intelligence for Strategic Agent-to-Agent Debt Recovery [65.30120701878582]
Large Language Model (LLM) agents are vulnerable to exploitation in emotion-sensitive domains like debt collection.<n>EmoDebt is an emotional intelligence engine that reframes a model's ability to express emotion in negotiation as a sequential decision-making problem.<n>EmoDebt achieves significant strategic robustness, substantially outperforming non-adaptive and emotion-agnostic baselines.
arXiv Detail & Related papers (2025-03-27T01:41:34Z) - EPO: Explicit Policy Optimization for Strategic Reasoning in LLMs via Reinforcement Learning [69.55982246413046]
We propose explicit policy optimization (EPO) for strategic reasoning.<n>We train the strategic reasoning model via multi-turn reinforcement learning (RL),utilizing process rewards and iterative self-play.<n>Our findings reveal various collaborative reasoning mechanisms emergent in EPO and its effectiveness in generating novel strategies.
arXiv Detail & Related papers (2025-02-18T03:15:55Z) - LLMs with Personalities in Multi-issue Negotiation Games [2.186901738997927]
We measure the ability of large language models (LLMs) to negotiate within a game-theoretical framework.
We find high openness, conscientiousness, and neuroticism are associated with fair tendencies.
Low agreeableness and low openness are associated with rational tendencies.
arXiv Detail & Related papers (2024-05-08T17:51:53Z) - K-Level Reasoning: Establishing Higher Order Beliefs in Large Language Models for Strategic Reasoning [76.3114831562989]
It requires Large Language Model (LLM) agents to adapt their strategies dynamically in multi-agent environments.
We propose a novel framework: "K-Level Reasoning with Large Language Models (K-R)"
arXiv Detail & Related papers (2024-02-02T16:07:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.