BattleAgent: Multi-modal Dynamic Emulation on Historical Battles to Complement Historical Analysis
- URL: http://arxiv.org/abs/2404.15532v1
- Date: Tue, 23 Apr 2024 21:37:22 GMT
- Title: BattleAgent: Multi-modal Dynamic Emulation on Historical Battles to Complement Historical Analysis
- Authors: Shuhang Lin, Wenyue Hua, Lingyao Li, Che-Jui Chang, Lizhou Fan, Jianchao Ji, Hang Hua, Mingyu Jin, Jiebo Luo, Yongfeng Zhang,
- Abstract summary: This paper presents BattleAgent, an emulation system that combines the Large Vision-Language Model and Multi-agent System.
It aims to simulate complex dynamic interactions among multiple agents, as well as between agents and their environments.
It emulates both the decision-making processes of leaders and the viewpoints of ordinary participants, such as soldiers.
- Score: 62.60458710368311
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents BattleAgent, an emulation system that combines the Large Vision-Language Model and Multi-agent System. This novel system aims to simulate complex dynamic interactions among multiple agents, as well as between agents and their environments, over a period of time. It emulates both the decision-making processes of leaders and the viewpoints of ordinary participants, such as soldiers. The emulation showcases the current capabilities of agents, featuring fine-grained multi-modal interactions between agents and landscapes. It develops customizable agent structures to meet specific situational requirements, for example, a variety of battle-related activities like scouting and trench digging. These components collaborate to recreate historical events in a lively and comprehensive manner while offering insights into the thoughts and feelings of individuals from diverse viewpoints. The technological foundations of BattleAgent establish detailed and immersive settings for historical battles, enabling individual agents to partake in, observe, and dynamically respond to evolving battle scenarios. This methodology holds the potential to substantially deepen our understanding of historical events, particularly through individual accounts. Such initiatives can also aid historical research, as conventional historical narratives often lack documentation and prioritize the perspectives of decision-makers, thereby overlooking the experiences of ordinary individuals. BattelAgent illustrates AI's potential to revitalize the human aspect in crucial social events, thereby fostering a more nuanced collective understanding and driving the progressive development of human society.
Related papers
- DialSim: A Real-Time Simulator for Evaluating Long-Term Dialogue Understanding of Conversational Agents [13.915753261117901]
We introduce DialSim, a real-time dialogue simulator.
In this simulator, an agent is assigned the role of a character from popular TV shows.
Key features of DialSim include evaluating the agent's ability to respond within a reasonable time limit.
arXiv Detail & Related papers (2024-06-19T01:37:10Z) - Hello Again! LLM-powered Personalized Agent for Long-term Dialogue [63.65128176360345]
We introduce a model-agnostic framework, the Long-term Dialogue Agent (LD-Agent)
It incorporates three independently tunable modules dedicated to event perception, persona extraction, and response generation.
The effectiveness, generality, and cross-domain capabilities of LD-Agent are empirically demonstrated.
arXiv Detail & Related papers (2024-06-09T21:58:32Z) - An Interactive Agent Foundation Model [49.77861810045509]
We propose an Interactive Agent Foundation Model that uses a novel multi-task agent training paradigm for training AI agents.
Our training paradigm unifies diverse pre-training strategies, including visual masked auto-encoders, language modeling, and next-action prediction.
We demonstrate the performance of our framework across three separate domains -- Robotics, Gaming AI, and Healthcare.
arXiv Detail & Related papers (2024-02-08T18:58:02Z) - MEIA: Towards Realistic Multimodal Interaction and Manipulation for Embodied Robots [82.67236400004826]
We introduce the Multimodal Embodied Interactive Agent (MEIA), capable of translating high-level tasks expressed in natural language into a sequence of executable actions.
MEM module enables MEIA to generate executable action plans based on diverse requirements and the robot's capabilities.
arXiv Detail & Related papers (2024-02-01T02:43:20Z) - Agent AI: Surveying the Horizons of Multimodal Interaction [83.18367129924997]
"Agent AI" is a class of interactive systems that can perceive visual stimuli, language inputs, and other environmentally-grounded data.
We envision a future where people can easily create any virtual reality or simulated scene and interact with agents embodied within the virtual environment.
arXiv Detail & Related papers (2024-01-07T19:11:18Z) - War and Peace (WarAgent): Large Language Model-based Multi-Agent
Simulation of World Wars [40.489161847202325]
We propose textbfWarAgent, an LLM-powered multi-agent AI system, to simulate historical international conflicts.
By evaluating the simulation effectiveness, we examine the advancements and limitations of cutting-edge AI systems' abilities.
Our findings offer data-driven and AI-augmented insights that can redefine how we approach conflict resolution and peacekeeping strategies.
arXiv Detail & Related papers (2023-11-28T20:59:49Z) - Generative Agents: Interactive Simulacra of Human Behavior [86.1026716646289]
We introduce generative agents--computational software agents that simulate believable human behavior.
We describe an architecture that extends a large language model to store a complete record of the agent's experiences.
We instantiate generative agents to populate an interactive sandbox environment inspired by The Sims.
arXiv Detail & Related papers (2023-04-07T01:55:19Z) - Multi-Agent Interplay in a Competitive Survival Environment [0.0]
This thesis is part of the author's thesis "Multi-Agent Interplay in a Competitive Survival Environment" for the Master's Degree in Artificial Intelligence and Robotics at Sapienza University of Rome, 2022.
arXiv Detail & Related papers (2023-01-19T12:04:03Z) - GATSBI: Generative Agent-centric Spatio-temporal Object Interaction [9.328991021103294]
GAT SBI is a generative model that transforms a sequence of raw observations into a structured representation.
We show GAT SBI achieves superior on scene decomposition and video prediction compared to its state-of-the-art counterparts.
arXiv Detail & Related papers (2021-04-09T09:45:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.