Automated Configuration of Negotiation Strategies
- URL: http://arxiv.org/abs/2004.00094v1
- Date: Tue, 31 Mar 2020 20:31:33 GMT
- Title: Automated Configuration of Negotiation Strategies
- Authors: Bram M. Renting (1), Holger H. Hoos (2), Catholijn M. Jonker (1 and 2)
((1) Delft University of Technology, (2) Leiden University)
- Abstract summary: Bidding and acceptance strategies have a substantial impact on the outcome of negotiations in scenarios with linear additive and nonlinear utility functions.
We develop a method leveraging automated algorithm configuration to find the best strategies for a specific set of negotiation settings.
We show that our automatically configured agent outperforms all other agents, with a 5.1% increase in negotiation payoff compared to the next-best agent.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bidding and acceptance strategies have a substantial impact on the outcome of
negotiations in scenarios with linear additive and nonlinear utility functions.
Over the years, it has become clear that there is no single best strategy for
all negotiation settings, yet many fixed strategies are still being developed.
We envision a shift in the strategy design question from: What is a good
strategy?, towards: What could be a good strategy? For this purpose, we
developed a method leveraging automated algorithm configuration to find the
best strategies for a specific set of negotiation settings. By empowering
automated negotiating agents using automated algorithm configuration, we obtain
a flexible negotiation agent that can be configured automatically for a rich
space of opponents and negotiation scenarios.
To critically assess our approach, the agent was tested in an ANAC-like
bilateral automated negotiation tournament setting against past competitors. We
show that our automatically configured agent outperforms all other agents, with
a 5.1% increase in negotiation payoff compared to the next-best agent. We note
that without our agent in the tournament, the top-ranked agent wins by a margin
of only 0.01%.
Related papers
- A Negotiator's Backup Plan: Optimal Concessions with a Reservation Value [3.3058382994863984]
This paper develops an optimal bidding strategy called MIA-RVelous for bilateral negotiations with private reservation values.
The proposed greedy algorithm finds the optimal bid sequence given the agent's beliefs about the opponent in $O(n2D)$ time.
arXiv Detail & Related papers (2024-04-30T08:45:18Z) - Strategic Apple Tasting [35.25249063553063]
Algorithmic decision-making in high-stakes domains often involves assigning decisions to agents with incentives to strategically modify their input to the algorithm.
We formalize this setting as an online learning problem with apple-tasting feedback.
Our goal is to achieve sublinear strategic regret, which compares the performance of the principal to that of the best fixed policy in hindsight.
arXiv Detail & Related papers (2023-06-09T20:46:31Z) - Provably Efficient Fictitious Play Policy Optimization for Zero-Sum
Markov Games with Structured Transitions [145.54544979467872]
We propose and analyze new fictitious play policy optimization algorithms for zero-sum Markov games with structured but unknown transitions.
We prove tight $widetildemathcalO(sqrtK)$ regret bounds after $K$ episodes in a two-agent competitive game scenario.
Our algorithms feature a combination of Upper Confidence Bound (UCB)-type optimism and fictitious play under the scope of simultaneous policy optimization.
arXiv Detail & Related papers (2022-07-25T18:29:16Z) - Decentralized Optimistic Hyperpolicy Mirror Descent: Provably No-Regret
Learning in Markov Games [95.10091348976779]
We study decentralized policy learning in Markov games where we control a single agent to play with nonstationary and possibly adversarial opponents.
We propose a new algorithm, underlineDecentralized underlineOptimistic hypeunderlineRpolicy munderlineIrror deunderlineScent (DORIS)
DORIS achieves $sqrtK$-regret in the context of general function approximation, where $K$ is the number of episodes.
arXiv Detail & Related papers (2022-06-03T14:18:05Z) - Conditional Imitation Learning for Multi-Agent Games [89.897635970366]
We study the problem of conditional multi-agent imitation learning, where we have access to joint trajectory demonstrations at training time.
We propose a novel approach to address the difficulties of scalability and data scarcity.
Our model learns a low-rank subspace over ego and partner agent strategies, then infers and adapts to a new partner strategy by interpolating in the subspace.
arXiv Detail & Related papers (2022-01-05T04:40:13Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - A Data-Driven Method for Recognizing Automated Negotiation Strategies [13.606307819976161]
We propose a novel data-driven approach for recognizing an opponent's s negotiation strategy.
Our approach includes a data generation method for an agent to generate domain-independent sequences.
We perform extensive experiments, spanning four problem scenarios, to demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2021-07-03T20:43:47Z) - A Cooperative-Competitive Multi-Agent Framework for Auto-bidding in
Online Advertising [53.636153252400945]
We propose a general Multi-Agent reinforcement learning framework for Auto-Bidding, namely MAAB, to learn the auto-bidding strategies.
Our approach outperforms several baseline methods in terms of social welfare and guarantees the ad platform's revenue.
arXiv Detail & Related papers (2021-06-11T08:07:14Z) - An Autonomous Negotiating Agent Framework with Reinforcement Learning
Based Strategies and Adaptive Strategy Switching Mechanism [3.4376560669160394]
This work focuses on solving the problem of expert selection and adapting to the opponent's behaviour with our Autonomous Negotiating Agent Framework.
Our framework has a reviewer component which enables self-enhancement capability by deciding to include new strategies or replace old ones with better strategies periodically.
arXiv Detail & Related papers (2021-02-06T14:38:03Z) - Universal Trading for Order Execution with Oracle Policy Distillation [99.57416828489568]
We propose a novel universal trading policy optimization framework to bridge the gap between the noisy yet imperfect market states and the optimal action sequences for order execution.
We show that our framework can better guide the learning of the common policy towards practically optimal execution by an oracle teacher with perfect information.
arXiv Detail & Related papers (2021-01-28T05:52:18Z) - A Deep Reinforcement Learning Approach to Concurrent Bilateral
Negotiation [6.484413431061962]
We present a novel negotiation model that allows an agent to learn how to negotiate during concurrent bilateral negotiations in unknown and dynamic e-markets.
The agent uses an actor-critic architecture with model-free reinforcement learning to learn a strategy expressed as a deep neural network.
As a result, we can build automated agents for concurrent negotiations that can adapt to different e-market settings without the need to be pre-programmed.
arXiv Detail & Related papers (2020-01-31T12:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.