Curious Causality-Seeking Agents Learn Meta Causal World
- URL: http://arxiv.org/abs/2506.23068v2
- Date: Fri, 01 Aug 2025 06:33:13 GMT
- Title: Curious Causality-Seeking Agents Learn Meta Causal World
- Authors: Zhiyu Zhao, Haoxuan Li, Haifeng Zhang, Jun Wang, Francesco Faccio, Jürgen Schmidhuber, Mengyue Yang,
- Abstract summary: We introduce the textbfMeta-Causal Graph as world models.<n>A single Meta-Causal Graph is composed of multiple causal subgraphs, each triggered by meta state.<n>Our method robustly captures shifts in causal dynamics and generalizes effectively to previously unseen contexts.
- Score: 33.20398438220672
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: When building a world model, a common assumption is that the environment has a single, unchanging underlying causal rule, like applying Newton's laws to every situation. In reality, what appears as a drifting causal mechanism is often the manifestation of a fixed underlying mechanism seen through a narrow observational window. This brings about a problem that, when building a world model, even subtle shifts in policy or environment states can alter the very observed causal mechanisms. In this work, we introduce the \textbf{Meta-Causal Graph} as world models, a minimal unified representation that efficiently encodes the transformation rules governing how causal structures shift across different latent world states. A single Meta-Causal Graph is composed of multiple causal subgraphs, each triggered by meta state, which is in the latent state space. Building on this representation, we introduce a \textbf{Causality-Seeking Agent} whose objectives are to (1) identify the meta states that trigger each subgraph, (2) discover the corresponding causal relationships by agent curiosity-driven intervention policy, and (3) iteratively refine the Meta-Causal Graph through ongoing curiosity-driven exploration and agent experiences. Experiments on both synthetic tasks and a challenging robot arm manipulation task demonstrate that our method robustly captures shifts in causal dynamics and generalizes effectively to previously unseen contexts.
Related papers
- SPARTAN: A Sparse Transformer Learning Local Causation [63.29645501232935]
Causal structures play a central role in world models that flexibly adapt to changes in the environment.
We present the SPARse TrANsformer World model (SPARTAN), a Transformer-based world model that learns local causal structures between entities in a scene.
By applying sparsity regularisation on the attention pattern between object-factored tokens, SPARTAN identifies sparse local causal models that accurately predict future object states.
arXiv Detail & Related papers (2024-11-11T11:42:48Z) - Failure Modes of LLMs for Causal Reasoning on Narratives [51.19592551510628]
We investigate the interaction between world knowledge and logical reasoning.<n>We find that state-of-the-art large language models (LLMs) often rely on superficial generalizations.<n>We show that simple reformulations of the task can elicit more robust reasoning behavior.
arXiv Detail & Related papers (2024-10-31T12:48:58Z) - Systems with Switching Causal Relations: A Meta-Causal Perspective [18.752058058199847]
flexibility of agents' actions or tipping points in the environmental process can change the qualitative dynamics of the system.<n>New causal relationships may emerge, while existing ones change or disappear, resulting in an altered causal graph.<n>We propose the concept of meta-causal states, which groups classical causal models into clusters based on equivalent qualitative behavior.
arXiv Detail & Related papers (2024-10-16T21:32:31Z) - Simplifying Latent Dynamics with Softly State-Invariant World Models [10.722955763425228]
We introduce the Parsimonious Latent Space Model (PLSM), a world model that regularizes the latent dynamics to make the effect of the agent's actions more predictable.
We find that our regularization improves accuracy, generalization, and performance in downstream tasks.
arXiv Detail & Related papers (2024-01-31T13:52:11Z) - Interpretable Imitation Learning with Dynamic Causal Relations [65.18456572421702]
We propose to expose captured knowledge in the form of a directed acyclic causal graph.
We also design this causal discovery process to be state-dependent, enabling it to model the dynamics in latent causal graphs.
The proposed framework is composed of three parts: a dynamic causal discovery module, a causality encoding module, and a prediction module, and is trained in an end-to-end manner.
arXiv Detail & Related papers (2023-09-30T20:59:42Z) - Causal Triplet: An Open Challenge for Intervention-centric Causal
Representation Learning [98.78136504619539]
Causal Triplet is a causal representation learning benchmark featuring visually more complex scenes.
We show that models built with the knowledge of disentangled or object-centric representations significantly outperform their distributed counterparts.
arXiv Detail & Related papers (2023-01-12T17:43:38Z) - DeGroot-based opinion formation under a global steering mechanism [3.0215418312651057]
We study the opinion formation process under the effect of a global steering mechanism (GSM), which aggregates the opinion-driven agent states at the network level.
We propose a new two-layer agent-based opinion formation model, called GSM-DeGroot, that captures the coupled dynamics between agent-to-agent local interactions.
arXiv Detail & Related papers (2022-10-21T22:02:47Z) - Towards a Grounded Theory of Causation for Embodied AI [12.259552039796027]
Existing frameworks give no indication as to which behaviour policies or physical transformations of state space shall count as interventions.
The framework sketched in this paper describes actions as transformations of state space, for instance induced by an agent running a policy.
This makes it possible to describe in a uniform way both transformations of the micro-state space and abstract models thereof.
arXiv Detail & Related papers (2022-06-28T12:56:43Z) - CausalCity: Complex Simulations with Agency for Causal Discovery and
Reasoning [68.74447489372037]
We present a high-fidelity simulation environment that is designed for developing algorithms for causal discovery and counterfactual reasoning.
A core component of our work is to introduce textitagency, such that it is simple to define and create complex scenarios.
We perform experiments with three state-of-the-art methods to create baselines and highlight the affordances of this environment.
arXiv Detail & Related papers (2021-06-25T00:21:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.