INTAGS: Interactive Agent-Guided Simulation
- URL: http://arxiv.org/abs/2309.01784v3
- Date: Fri, 17 Nov 2023 20:05:46 GMT
- Title: INTAGS: Interactive Agent-Guided Simulation
- Authors: Song Wei, Andrea Coletta, Svitlana Vyetrenko, Tucker Balch
- Abstract summary: In many applications involving multi-agent system (MAS), it is imperative to test an experimental (Exp) autonomous agent in a high-fidelity simulator prior to its deployment to production.
We propose a metric to distinguish between real and synthetic multi-agent systems, which is evaluated through the live interaction between the Exp and BG agents.
We show that using INTAGS to calibrate the simulator can generate more realistic market data compared to the state-of-the-art conditional Wasserstein Generative Adversarial Network approach.
- Score: 4.04638613278729
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In many applications involving multi-agent system (MAS), it is imperative to
test an experimental (Exp) autonomous agent in a high-fidelity simulator prior
to its deployment to production, to avoid unexpected losses in the real-world.
Such a simulator acts as the environmental background (BG) agent(s), called
agent-based simulator (ABS), aiming to replicate the complex real MAS. However,
developing realistic ABS remains challenging, mainly due to the sequential and
dynamic nature of such systems. To fill this gap, we propose a metric to
distinguish between real and synthetic multi-agent systems, which is evaluated
through the live interaction between the Exp and BG agents to explicitly
account for the systems' sequential nature. Specifically, we characterize the
system/environment by studying the effect of a sequence of BG agents' responses
to the environment state evolution and take such effects' differences as MAS
distance metric; The effect estimation is cast as a causal inference problem
since the environment evolution is confounded with the previous environment
state. Importantly, we propose the Interactive Agent-Guided Simulation (INTAGS)
framework to build a realistic ABS by optimizing over this novel metric. To
adapt to any environment with interactive sequential decision making agents,
INTAGS formulates the simulator as a stochastic policy in reinforcement
learning. Moreover, INTAGS utilizes the policy gradient update to bypass
differentiating the proposed metric such that it can support non-differentiable
operations of multi-agent environments. Through extensive experiments, we
demonstrate the effectiveness of INTAGS on an equity stock market simulation
example. We show that using INTAGS to calibrate the simulator can generate more
realistic market data compared to the state-of-the-art conditional Wasserstein
Generative Adversarial Network approach.
Related papers
- Enabling Adaptive Agent Training in Open-Ended Simulators by Targeting Diversity [10.402855891273346]
DIVA is an evolutionary approach for generating diverse training tasks in complex, open-ended simulators.
Our empirical results showcase DIVA's unique ability to overcome complex parameterizations and successfully train adaptive agent behavior.
arXiv Detail & Related papers (2024-11-07T06:27:12Z) - A Simulation System Towards Solving Societal-Scale Manipulation [14.799498804818333]
The rise of AI-driven manipulation poses significant risks to societal trust and democratic processes.
Yet, studying these effects in real-world settings at scale is ethically and logistically impractical.
We present a simulation environment designed to address this.
arXiv Detail & Related papers (2024-10-17T03:16:24Z) - Evaluating Real-World Robot Manipulation Policies in Simulation [91.55267186958892]
Control and visual disparities between real and simulated environments are key challenges for reliable simulated evaluation.
We propose approaches for mitigating these gaps without needing to craft full-fidelity digital twins of real-world environments.
We create SIMPLER, a collection of simulated environments for manipulation policy evaluation on common real robot setups.
arXiv Detail & Related papers (2024-05-09T17:30:16Z) - S2R-ViT for Multi-Agent Cooperative Perception: Bridging the Gap from
Simulation to Reality [41.25312194294171]
We propose the first-to-Reality transfer learning framework for multi-agent cooperative perception using a novel Vision Transformer, named as S2R-ViT.
Our experiments on the public multi-agent cooperative perception datasets OPV2V and V2V4Real demonstrate that the proposed S2R-ViT can effectively bridge the gap from simulation to reality.
arXiv Detail & Related papers (2023-07-16T03:54:10Z) - Differentiable Agent-based Epidemiology [71.81552021144589]
We introduce GradABM: a scalable, differentiable design for agent-based modeling that is amenable to gradient-based learning with automatic differentiation.
GradABM can quickly simulate million-size populations in few seconds on commodity hardware, integrate with deep neural networks and ingest heterogeneous data sources.
arXiv Detail & Related papers (2022-07-20T07:32:02Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - Nonprehensile Riemannian Motion Predictive Control [57.295751294224765]
We introduce a novel Real-to-Sim reward analysis technique to reliably imagine and predict the outcome of taking possible actions for a real robotic platform.
We produce a closed-loop controller to reactively push objects in a continuous action space.
We observe that RMPC is robust in cluttered as well as occluded environments and outperforms the baselines.
arXiv Detail & Related papers (2021-11-15T18:50:04Z) - ERMAS: Becoming Robust to Reward Function Sim-to-Real Gaps in
Multi-Agent Simulations [110.72725220033983]
Epsilon-Robust Multi-Agent Simulation (ERMAS) is a framework for learning AI policies that are robust to such multiagent sim-to-real gaps.
ERMAS learns tax policies that are robust to changes in agent risk aversion, improving social welfare by up to 15% in complextemporal simulations.
In particular, ERMAS learns tax policies that are robust to changes in agent risk aversion, improving social welfare by up to 15% in complextemporal simulations.
arXiv Detail & Related papers (2021-06-10T04:32:20Z) - TrafficSim: Learning to Simulate Realistic Multi-Agent Behaviors [74.67698916175614]
We propose TrafficSim, a multi-agent behavior model for realistic traffic simulation.
In particular, we leverage an implicit latent variable model to parameterize a joint actor policy.
We show TrafficSim generates significantly more realistic and diverse traffic scenarios as compared to a diverse set of baselines.
arXiv Detail & Related papers (2021-01-17T00:29:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.