Active Inverse Learning in Stackelberg Trajectory Games
- URL: http://arxiv.org/abs/2308.08017v1
- Date: Tue, 15 Aug 2023 20:17:26 GMT
- Title: Active Inverse Learning in Stackelberg Trajectory Games
- Authors: Yue Yu, Jacob Levy, Negar Mehr, David Fridovich-Keil, and Ufuk Topcu
- Abstract summary: We formulate an inverse learning problem in a Stackelberg game between a leader and a follower.
We propose an active inverse learning method for the leader to infer which hypothesis among a finite set of candidates describes the follower's objective function.
- Score: 34.6111453235322
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Game-theoretic inverse learning is the problem of inferring the players'
objectives from their actions. We formulate an inverse learning problem in a
Stackelberg game between a leader and a follower, where each player's action is
the trajectory of a dynamical system. We propose an active inverse learning
method for the leader to infer which hypothesis among a finite set of
candidates describes the follower's objective function. Instead of using
passively observed trajectories like existing methods, the proposed method
actively maximizes the differences in the follower's trajectories under
different hypotheses to accelerate the leader's inference. We demonstrate the
proposed method in a receding-horizon repeated trajectory game. Compared with
uniformly random inputs, the leader inputs provided by the proposed method
accelerate the convergence of the probability of different hypotheses
conditioned on the follower's trajectory by orders of magnitude.
Related papers
- Clustering Human Mobility with Multiple Spaces [8.076957089365676]
This paper proposes a novel mobility clustering method for mobility behavior detection.
The proposed method uses a variational autoencoder architecture to simultaneously perform clustering in both latent and original spaces.
The experiment shows that the proposed method outperformed state-of-the-art methods in mobility behavior detection from trajectories with better accuracy and more interpretability.
arXiv Detail & Related papers (2023-01-20T12:02:30Z) - No-Regret Learning in Dynamic Stackelberg Games [31.001205916012307]
In a Stackelberg game, a leader commits to a randomized strategy, and a follower chooses their best strategy in response.
We consider an extension of a standard Stackelberg game, called a discrete-time dynamic Stackelberg game, that has an underlying state space that affects the leader's rewards and available strategies and evolves in a Markovian manner depending on both the leader and follower's selected strategies.
arXiv Detail & Related papers (2022-02-10T01:07:57Z) - Human Trajectory Prediction via Counterfactual Analysis [87.67252000158601]
Forecasting human trajectories in complex dynamic environments plays a critical role in autonomous vehicles and intelligent robots.
Most existing methods learn to predict future trajectories by behavior clues from history trajectories and interaction clues from environments.
We propose a counterfactual analysis method for human trajectory prediction to investigate the causality between the predicted trajectories and input clues.
arXiv Detail & Related papers (2021-07-29T17:41:34Z) - Learning to drive from a world on rails [78.28647825246472]
We learn an interactive vision-based driving policy from pre-recorded driving logs via a model-based approach.
A forward model of the world supervises a driving policy that predicts the outcome of any potential driving trajectory.
Our method ranks first on the CARLA leaderboard, attaining a 25% higher driving score while using 40 times less data.
arXiv Detail & Related papers (2021-05-03T05:55:30Z) - Adversarial Training as Stackelberg Game: An Unrolled Optimization
Approach [91.74682538906691]
Adversarial training has been shown to improve the generalization performance of deep learning models.
We propose Stackelberg Adversarial Training (SALT), which formulates adversarial training as a Stackelberg game.
arXiv Detail & Related papers (2021-04-11T00:44:57Z) - Learning to Shift Attention for Motion Generation [55.61994201686024]
One challenge of motion generation using robot learning from demonstration techniques is that human demonstrations follow a distribution with multiple modes for one task query.
Previous approaches fail to capture all modes or tend to average modes of the demonstrations and thus generate invalid trajectories.
We propose a motion generation model with extrapolation ability to overcome this problem.
arXiv Detail & Related papers (2021-02-24T09:07:52Z) - Reward Conditioned Neural Movement Primitives for Population Based
Variational Policy Optimization [4.559353193715442]
This paper studies the reward based policy exploration problem in a supervised learning approach.
We show that our method provides stable learning progress and significant sample efficiency compared to a number of state-of-the-art robotic reinforcement learning methods.
arXiv Detail & Related papers (2020-11-09T09:53:37Z) - Learning Game-Theoretic Models of Multiagent Trajectories Using Implicit
Layers [9.594432031144716]
We propose an end-to-end trainable architecture that hybridizes neural nets with game-theoretic reasoning.
For tractability, we introduce a new class of continuous potential games and an equilibrium-separating partition of the action space.
We evaluate our approach on two real-world data sets, where we predict highway merging driver trajectories, and on a simple decision-making transfer task.
arXiv Detail & Related papers (2020-08-17T13:34:12Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.