Asymptotic Semantic Collapse in Hierarchical Optimization
- URL: http://arxiv.org/abs/2602.18450v1
- Date: Sun, 01 Feb 2026 00:02:01 GMT
- Title: Asymptotic Semantic Collapse in Hierarchical Optimization
- Authors: Faruk Alpay, Bugra Kilictas,
- Abstract summary: Multi-agent language systems can exhibit a failure mode where a shared dominant context progressively absorbs individual semantics.<n>We study this effect under the name Asymptotic Semantic Collapse in Hierarchical Optimization.<n>We show that repeated interactions with Peripheral Agent Nodes drive an alignment that minimizes a global loss.
- Score: 0.5729426778193398
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-agent language systems can exhibit a failure mode where a shared dominant context progressively absorbs individual semantics, yielding near-uniform behavior across agents. We study this effect under the name Asymptotic Semantic Collapse in Hierarchical Optimization. In a closed linguistic setting with a Dominant Anchor Node whose semantic state has effectively infinite inertia, we show that repeated interactions with Peripheral Agent Nodes drive an asymptotic alignment that minimizes a global loss. We model semantic states as points on a Riemannian manifold and analyze the induced projection dynamics. Two consequences follow. First, the limiting semantic configuration is insensitive to the optimization history: both smooth gradient-style updates and stochastic noisy updates converge to the same topological endpoint, establishing path independence at convergence. Second, the degree of context dependence controls information content: moving from atomic (independent) representations to fully entangled (context-bound) representations forces the node entropy, interpreted as available degrees of freedom, to vanish in the limit. The theory connects information-theoretic quantities with differential-geometric structure and suggests an interpretation as an immutable consensus rule that constrains agents to a shared semantic grammar. A lightweight dataset-free benchmark on an RWKV-7 13B GGUF checkpoint complements the analysis, reporting zero hash collisions, mean compliance of 0.50 under greedy decoding and 0.531 under stochastic decoding, and final Jaccard-to-anchor similarity values of 0.295 and 0.224, respectively.
Related papers
- Riemannian Flow Matching for Disentangled Graph Domain Adaptation [51.98961391065951]
Graph Domain Adaptation (GDA) typically uses adversarial learning to align graph embeddings in Euclidean space.<n>DisRFM is a geometry-aware GDA framework that unifies embedding and flow-based transport.
arXiv Detail & Related papers (2026-01-31T11:05:35Z) - ICON: Invariant Counterfactual Optimization with Neuro-Symbolic Priors for Text-Based Person Search [6.247167721048087]
Text-Based Person Search holds unique value in real-world surveillance bridging visual perception and language understanding.<n>Current paradigms utilizing pre-training models often fail to transfer effectively to complex open-world scenarios.<n>This paper proposes ICON, a framework integrating causal and topological priors.
arXiv Detail & Related papers (2026-01-22T13:09:22Z) - A Foundational Theory of Quantitative Abstraction: Adjunctions, Duality, and Logic for Probabilistic Systems [2.362412515574206]
Large or continuous state spaces make exact analysis intractable and call for principled quantitative abstraction.<n>This work develops a unified theory of such abstraction by integrating category theory, coalgebra, quantitative logic, and optimal transport.
arXiv Detail & Related papers (2025-10-22T10:16:24Z) - Pure Exploration via Frank-Wolfe Self-Play [9.025261095806309]
We study pure exploration in structured multi-armed bandits, aiming to efficiently identify the correct hypothesis from a finite set of alternatives.<n>Our analysis proceeds through a continuous-time steer: a differential inclusion with a Lyapunov function that decays, implying a vanishing duality gap and convergence to the optimal value.
arXiv Detail & Related papers (2025-09-24T08:55:21Z) - Semantic Loss Functions for Neuro-Symbolic Structured Prediction [74.18322585177832]
We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training.
It is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby.
It can be combined with both discriminative and generative neural models.
arXiv Detail & Related papers (2024-05-12T22:18:25Z) - Federated Learning Resilient to Byzantine Attacks and Data Heterogeneity [59.17297282373628]
This paper addresses Gradient learning (FL) in the context of malicious attacks on data.<n>We introduce a novel Average Robust Algorithm (RAGA) which uses the median for both convergence analysis and loss functions.
arXiv Detail & Related papers (2024-03-20T08:15:08Z) - STAR Loss: Reducing Semantic Ambiguity in Facial Landmark Detection [80.04000067312428]
We propose a Self-adapTive Ambiguity Reduction (STAR) loss by exploiting the properties of semantic ambiguity.
We find that semantic ambiguity results in the anisotropic predicted distribution, which inspires us to use predicted distribution to represent semantic ambiguity.
We also propose two kinds of eigenvalue restriction methods that could avoid both distribution's abnormal change and the model's premature convergence.
arXiv Detail & Related papers (2023-06-05T10:33:25Z) - Convergence of Adam Under Relaxed Assumptions [72.24779199744954]
We show that Adam converges to $epsilon$-stationary points with $O(epsilon-4)$ gradient complexity under far more realistic conditions.
We also propose a variance-reduced version of Adam with an accelerated gradient complexity of $O(epsilon-3)$.
arXiv Detail & Related papers (2023-04-27T06:27:37Z) - The distribution of syntactic dependency distances [0.13812010983144798]
We contribute to the characterization of the actual distribution of syntactic dependency distances.<n>We propose a new model with two exponential regimes in which the probability decay is allowed to change after a break-point.<n>We find that a two-regime model is the most likely one in all 20 languages we considered, independently of sentence length and annotation style.
arXiv Detail & Related papers (2022-11-26T17:31:25Z) - Gradient flows and randomised thresholding: sparse inversion and
classification [0.0]
Sparse inversion and classification problems are ubiquitous in modern data science and imaging.
In classification, we consider, e.g., the sum of a data fidelity term and a non-smooth Ginzburg--Landau energy.
Standard (sub)gradient descent methods have shown to be inefficient when approaching such problems.
arXiv Detail & Related papers (2022-03-22T09:21:14Z) - Neuro-Symbolic Entropy Regularization [78.16196949641079]
In structured prediction, the goal is to jointly predict many output variables that together encode a structured object.
One approach -- entropy regularization -- posits that decision boundaries should lie in low-probability regions.
We propose a loss, neuro-symbolic entropy regularization, that encourages the model to confidently predict a valid object.
arXiv Detail & Related papers (2022-01-25T06:23:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.