Towards Error-Centric Intelligence II: Energy-Structured Causal Models
- URL: http://arxiv.org/abs/2510.22050v1
- Date: Fri, 24 Oct 2025 22:19:17 GMT
- Title: Towards Error-Centric Intelligence II: Energy-Structured Causal Models
- Authors: Marcus Thomas,
- Abstract summary: We argue for a conceptual reorientation: intelligence is the ability to build and refine explanations.<n>This paper offers a formal language for causal reasoning in systems that aspire to understand, not merely to predict.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Contemporary machine learning optimizes for predictive accuracy, yet systems that achieve state of the art performance remain causally opaque: their internal representations provide no principled handle for intervention. We can retrain such models, but we cannot surgically edit specific mechanisms while holding others fixed, because learned latent variables lack causal semantics. We argue for a conceptual reorientation: intelligence is the ability to build and refine explanations, falsifiable claims about manipulable structure that specify what changes and what remains invariant under intervention. Explanations subsume prediction but demand more: causal commitments that can be independently tested and corrected at the level of mechanisms. We introduce computational explanations, mappings from observations to intervention ready causal accounts. We instantiate these explanations with Energy Structured Causal Models (ESCMs), in which mechanisms are expressed as constraints (energy functions or vector fields) rather than explicit input output maps, and interventions act by local surgery on those constraints. This shift makes internal structure manipulable at the level where explanations live: which relations must hold, which can change, and what follows when they do. We provide concrete instantiations of the structural-causal principles LAP and ICM in the ESCM context, and also argue that empirical risk minimization systematically produces fractured, entangled representations, a failure we analyze as gauge ambiguity in encoder energy pairs. Finally, we show that under mild conditions, ESCMs recover standard SCM semantics. Building on Part I's principles (LAP, ICM, CAP) and its definition of intelligence as explanation-building under criticism, this paper offers a formal language for causal reasoning in systems that aspire to understand, not merely to predict.
Related papers
- Subject-Event Ontology Without Global Time: Foundations and Execution Semantics [51.56484100374058]
The formalization includes nine axioms (A1-A9), ensuring the correctness of executable: monotonicity of history (I1), acyclicity of causality (I2), traceability (I3)<n>The formalization is applicable to distributed systems, microservice architectures, DLT platforms, and multiperspectivity scenarios (conflicting facts from different subjects)<n>Special attention is given to the model-based approach (A9): event validation via schemas, actor authorization, automatic construction of causal chains (W3) without global time.
arXiv Detail & Related papers (2025-10-20T19:26:44Z) - Towards Error Centric Intelligence I, Beyond Observational Learning [0.0]
We argue that progress toward AGI is theory limited rather than data or scale limited.<n>We begin by laying foundations, definitions of knowledge, learning, intelligence, counterfactual competence and AGI.<n>We recast the problem as three questions about how explicit and implicit errors evolve under an agent's actions.
arXiv Detail & Related papers (2025-10-16T20:33:55Z) - Causal Reflection with Language Models [0.08594140167290097]
We introduce Causal Reflection, a framework that explicitly models causality as a dynamic function over state, action, time, and perturbation.<n>We define a formal Reflect mechanism that identifies mismatches between predicted and observed outcomes and generates causal hypotheses to revise the agent's internal model.<n>Our framework lays the theoretical groundwork for Causal Reflective agents that can adapt, self-correct, and communicate causal understanding in evolving environments.
arXiv Detail & Related papers (2025-08-06T14:44:23Z) - Comprehension Without Competence: Architectural Limits of LLMs in Symbolic Computation and Reasoning [8.248971739035746]
Large Language Models (LLMs) display striking surface fluency yet systematically fail at tasks requiring symbolic reasoning, arithmetic accuracy, and logical consistency.<n>This paper offers a structural diagnosis of such failures, revealing a persistent gap between textitcomprehension and textitcompetence<n>We term this phenomenon the computational textitsplit-brain syndrome, where instruction and action pathways are geometrically and functionally dissociated.
arXiv Detail & Related papers (2025-07-14T04:01:45Z) - Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers [76.42159902257677]
We argue that both behaviors stem from a single mechanism known as out-of-context reasoning (OCR)<n>OCR drives both generalization and hallucination, depending on whether the associated concepts are causally related.<n>Our work provides a theoretical foundation for understanding the OCR phenomenon, offering a new lens for analyzing and mitigating undesirable behaviors from knowledge injection.
arXiv Detail & Related papers (2025-06-12T16:50:45Z) - How do Transformers Learn Implicit Reasoning? [67.02072851088637]
We study how implicit multi-hop reasoning emerges by training transformers from scratch in a controlled symbolic environment.<n>We find that training with atomic triples is not necessary but accelerates learning, and that second-hop generalization relies on query-level exposure to specific compositional structures.
arXiv Detail & Related papers (2025-05-29T17:02:49Z) - Decomposing Interventional Causality into Synergistic, Redundant, and Unique Components [0.0]
We introduce a novel framework for decomposing interventional causal effects into synergistic, redundant, and unique components.<n>We develop a mathematical approach that systematically quantifies how causal power is distributed among variables in a system.
arXiv Detail & Related papers (2025-01-20T12:34:51Z) - Self-Healing Machine Learning: A Framework for Autonomous Adaptation in Real-World Environments [50.310636905746975]
Real-world machine learning systems often encounter model performance degradation due to distributional shifts in the underlying data generating process.
Existing approaches to addressing shifts, such as concept drift adaptation, are limited by their reason-agnostic nature.
We propose self-healing machine learning (SHML) to overcome these limitations.
arXiv Detail & Related papers (2024-10-31T20:05:51Z) - Diagnosing and Rectifying Fake OOD Invariance: A Restructured Causal
Approach [51.012396632595554]
Invariant representation learning (IRL) encourages the prediction from invariant causal features to labels de-confounded from the environments.
Recent theoretical results verified that some causal features recovered by IRLs merely pretend domain-invariantly in the training environments but fail in unseen domains.
We develop an approach based on conditional mutual information with respect to RS-SCM, then rigorously rectify the spurious and fake invariant effects.
arXiv Detail & Related papers (2023-12-15T12:58:05Z) - Towards a Grounded Theory of Causation for Embodied AI [12.259552039796027]
Existing frameworks give no indication as to which behaviour policies or physical transformations of state space shall count as interventions.
The framework sketched in this paper describes actions as transformations of state space, for instance induced by an agent running a policy.
This makes it possible to describe in a uniform way both transformations of the micro-state space and abstract models thereof.
arXiv Detail & Related papers (2022-06-28T12:56:43Z) - Discovering Latent Causal Variables via Mechanism Sparsity: A New
Principle for Nonlinear ICA [81.4991350761909]
Independent component analysis (ICA) refers to an ensemble of methods which formalize this goal and provide estimation procedure for practical application.
We show that the latent variables can be recovered up to a permutation if one regularizes the latent mechanisms to be sparse.
arXiv Detail & Related papers (2021-07-21T14:22:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.