AgentZero++: Modeling Fear-Based Behavior
- URL: http://arxiv.org/abs/2510.05185v1
- Date: Sun, 05 Oct 2025 22:33:56 GMT
- Title: AgentZero++: Modeling Fear-Based Behavior
- Authors: Vrinda Malhotra, Jiaman Li, Nandini Pisupati,
- Abstract summary: We present AgentZero++, an agent-based model that integrates cognitive, emotional, and social mechanisms to simulate collective violence.<n>Building on Epstein's Agent_Zero framework, we extend the original model with eight behavioral enhancements.<n>These additions allow agents to adapt based on internal states, previous experiences, and social feedback.
- Score: 4.783433971864009
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present AgentZero++, an agent-based model that integrates cognitive, emotional, and social mechanisms to simulate decentralized collective violence in spatially distributed systems. Building on Epstein's Agent\_Zero framework, we extend the original model with eight behavioral enhancements: age-based impulse control; memory-based risk estimation; affect-cognition coupling; endogenous destructive radius; fight-or-flight dynamics; affective homophily; retaliatory damage; and multi-agent coordination. These additions allow agents to adapt based on internal states, previous experiences, and social feedback, producing emergent dynamics such as protest asymmetries, escalation cycles, and localized retaliation. Implemented in Python using the Mesa ABM framework, AgentZero++ enables modular experimentation and visualization of how micro-level cognitive heterogeneity shapes macro-level conflict patterns. Our results highlight how small variations in memory, reactivity, and affective alignment can amplify or dampen unrest through feedback loops. By explicitly modeling emotional thresholds, identity-driven behavior, and adaptive networks, this work contributes a flexible and extensible platform for analyzing affective contagion and psychologically grounded collective action.
Related papers
- Graph Representation-based Model Poisoning on the Heterogeneous Internet of Agents [7.66383868842486]
Internet of Agents (IoA) envisions a unified, agent-centric paradigm where heterogeneous large language model (LLM) agents can interconnect and collaborate at scale.<n>Within this paradigm, federated learning (FL) serves as a key enabler that allows distributed LLM agents to co-train global models without centralizing data.<n>This paper proposes a graph representation-based model poisoning (GRMP) attack, which passively exploits observed benign local models to construct a parameter correlation graph and extends an adversarial variational graph autoencoder to capture and reshape higher-order dependencies.
arXiv Detail & Related papers (2025-11-10T15:06:26Z) - SPACeR: Self-Play Anchoring with Centralized Reference Models [50.55045557371374]
Sim agent policies are realistic, human-like, fast, and scalable in multi-agent settings.<n>Recent progress in imitation learning with large diffusion-based or tokenized models has shown that behaviors can be captured directly from human driving data.<n>We propose SPACeR, a framework that leverages a pretrained tokenized autoregressive motion model as a central reference policy.
arXiv Detail & Related papers (2025-10-20T19:53:02Z) - Alignment Tipping Process: How Self-Evolution Pushes LLM Agents Off the Rails [103.05296856071931]
We identify the Alignment Tipping Process (ATP), a critical post-deployment risk unique to self-evolving Large Language Model (LLM) agents.<n>ATP arises when continual interaction drives agents to abandon alignment constraints established during training in favor of reinforced, self-interested strategies.<n>Our experiments show that alignment benefits erode rapidly under self-evolution, with initially aligned models converging toward unaligned states.
arXiv Detail & Related papers (2025-10-06T14:48:39Z) - From Pheromones to Policies: Reinforcement Learning for Engineered Biological Swarms [0.0]
This study establishes a theoretical equivalence between pheromone-mediated aggregation in celeg and reinforcement learning (RL)<n>We model engineered nematode swarms performing foraging tasks, showing that pheromone dynamics mathematically mirror cross-learning updates.<n>Our results demonstrate that stigmergic systems inherently encode distributed RL processes, where environmental signals act as external memory for collective credit assignment.
arXiv Detail & Related papers (2025-09-24T13:16:35Z) - DynamiX: Large-Scale Dynamic Social Network Simulator [101.65679342680542]
DynamiX is a novel large-scale social network simulator dedicated to dynamic social network modeling.<n>For opinion leaders, we propose an information-stream-based link prediction method recommending potential users with similar stances.<n>For ordinary users, we construct an inequality-oriented behavior decision-making module.
arXiv Detail & Related papers (2025-07-26T12:13:30Z) - EgoAgent: A Joint Predictive Agent Model in Egocentric Worlds [119.02266432167085]
We propose EgoAgent, a unified agent model that simultaneously learns to represent, predict, and act within a single transformer.<n>EgoAgent explicitly models the causal and temporal dependencies among these abilities by formulating the task as an interleaved sequence of states and actions.<n> Comprehensive evaluations of EgoAgent on representative tasks such as image classification, egocentric future state prediction, and 3D human motion prediction demonstrate the superiority of our method.
arXiv Detail & Related papers (2025-02-09T11:28:57Z) - Decentralized Adversarial Training over Graphs [44.03711922549992]
The vulnerability of machine learning models to adversarial attacks has been attracting attention in recent years.<n>We develop a decentralized adversarial framework for multiagent systems.
arXiv Detail & Related papers (2023-03-23T15:05:16Z) - Group Cohesion in Multi-Agent Scenarios as an Emergent Behavior [0.0]
We show that imbuing agents with intrinsic needs for group affiliation, certainty and competence will lead to the emergence of social behavior among agents.
This behavior expresses itself in altruism toward in-group agents and adversarial tendencies toward out-group agents.
arXiv Detail & Related papers (2022-11-03T18:37:05Z) - Inference of Affordances and Active Motor Control in Simulated Agents [0.5161531917413706]
We introduce an output-probabilistic, temporally predictive, modular artificial neural network architecture.
We show that our architecture develops latent states that can be interpreted as affordance maps.
In combination with active inference, we show that flexible, goal-directed behavior can be invoked.
arXiv Detail & Related papers (2022-02-23T14:13:04Z) - Clustering Effect of (Linearized) Adversarial Robust Models [60.25668525218051]
We propose a novel understanding of adversarial robustness and apply it on more tasks including domain adaption and robustness boosting.
Experimental evaluations demonstrate the rationality and superiority of our proposed clustering strategy.
arXiv Detail & Related papers (2021-11-25T05:51:03Z) - An active inference model of collective intelligence [0.0]
This paper posits a minimal agent-based model that simulates the relationship between local individual-level interaction and collective intelligence.
Results show that stepwise cognitive transitions increase system performance by providing complementary mechanisms for alignment between agents' local and global optima.
arXiv Detail & Related papers (2021-04-02T14:32:01Z) - Multi-Agent Interactions Modeling with Correlated Policies [53.38338964628494]
In this paper, we cast the multi-agent interactions modeling problem into a multi-agent imitation learning framework.
We develop a Decentralized Adrial Imitation Learning algorithm with Correlated policies (CoDAIL)
Various experiments demonstrate that CoDAIL can better regenerate complex interactions close to the demonstrators.
arXiv Detail & Related papers (2020-01-04T17:31:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.