From Model Design to Organizational Design: Complexity Redistribution and Trade-Offs in Generative AI
- URL: http://arxiv.org/abs/2506.22440v1
- Date: Tue, 10 Jun 2025 15:22:09 GMT
- Title: From Model Design to Organizational Design: Complexity Redistribution and Trade-Offs in Generative AI
- Authors: Sharique Hasan, Alexander Oettl, Sampsa Samila,
- Abstract summary: We argue that viewing AI as a simple reduction in input costs overlooks two critical dynamics.<n>The GAS trade-off, therefore, does not disappear but is relocated from the user to the organization.<n>This study advances AI strategy by clarifying how scalable cognition relocates complexity.
- Score: 44.99833362998488
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces the Generality-Accuracy-Simplicity (GAS) framework to analyze how large language models (LLMs) are reshaping organizations and competitive strategy. We argue that viewing AI as a simple reduction in input costs overlooks two critical dynamics: (a) the inherent trade-offs among generality, accuracy, and simplicity, and (b) the redistribution of complexity across stakeholders. While LLMs appear to defy the traditional trade-off by offering high generality and accuracy through simple interfaces, this user-facing simplicity masks a significant shift of complexity to infrastructure, compliance, and specialized personnel. The GAS trade-off, therefore, does not disappear but is relocated from the user to the organization, creating new managerial challenges, particularly around accuracy in high-stakes applications. We contend that competitive advantage no longer stems from mere AI adoption, but from mastering this redistributed complexity through the design of abstraction layers, workflow alignment, and complementary expertise. This study advances AI strategy by clarifying how scalable cognition relocates complexity and redefines the conditions for technology integration.
Related papers
- Agentic Adversarial QA for Improving Domain-Specific LLMs [53.00642389531106]
Large Language Models (LLMs) often struggle to adapt effectively to specialized domains.<n>We propose an adversarial question-generation framework that produces a compact set of semantically challenging questions.
arXiv Detail & Related papers (2026-02-20T10:53:09Z) - Agentic Proposing: Enhancing Large Language Model Reasoning via Compositional Skill Synthesis [10.951981109673119]
Agentic Proposing is a framework that models problem synthesis as a goal-driven sequential decision process.<n>It generates high-precision, verifiable training trajectories across mathematics, coding, and science.<n>A 30B solver trained on only 11,000 synthesized trajectories achieves a state-of-the-art 91.6% accuracy on AIME25.
arXiv Detail & Related papers (2026-02-03T09:02:53Z) - Plain Transformers are Surprisingly Powerful Link Predictors [57.01966734467712]
Link prediction is a core challenge in graph machine learning, demanding models that capture rich and complex topological dependencies.<n>While Graph Neural Networks (GNNs) are the standard solution, state-of-the-art pipelines often rely on explicit structurals or memory-intensive node embeddings.<n>We present PENCIL, an encoder-only plain Transformer that replaces hand-crafted priors with attention over sampled local subgraphs.
arXiv Detail & Related papers (2026-02-02T02:45:52Z) - Simplifying Multi-Task Architectures Through Task-Specific Normalization [0.9668407688201359]
Multi-task learning (MTL) aims to leverage shared knowledge across tasks to improve generalization and parameter efficiency.<n>We show that normalization layers alone are sufficient to address many of these challenges.<n>We propose Task-Specific Sigmoid Batch Normalization (TS$$BN), a lightweight mechanism that enables tasks to softly allocate network capacity.
arXiv Detail & Related papers (2025-12-23T15:02:12Z) - ORPR: An OR-Guided Pretrain-then-Reinforce Learning Model for Inventory Management [9.138155308817215]
"Pretrain-then-Reinforce" approach reconciles AI's adaptive perception with Operations Research's structural rigor.<n>We show that a lightweight, domain-informed model can deliver state-of-the-art performance and robust transferability when guided by structured OR logic.
arXiv Detail & Related papers (2025-12-22T03:39:43Z) - Rethinking Multi-Agent Intelligence Through the Lens of Small-World Networks [14.233668486426795]
Large language models (LLMs) have enabled multi-agent systems (MAS) in which multiple agents argue, critique, and coordinate to solve complex tasks.<n>Most existing LLM-based MAS either adopt fully connected graphs, simple sparse rings, or ad-hoc dynamic selection, with little structural guidance.<n>We first bridge insights from neuroscience and complex networks to MAS, highlighting how SW structures balance local clustering and long-range integration.<n>Experiment results show that SW connectivity yields nearly the same accuracy and token cost, while substantially stabilizing consensus trajectories.
arXiv Detail & Related papers (2025-12-19T22:05:43Z) - Dynamic Generation of Multi-LLM Agents Communication Topologies with Graph Diffusion Models [99.85131798240808]
We introduce a novel generative framework called textitGuided Topology Diffusion (GTD)<n>Inspired by conditional discrete graph diffusion models, GTD formulates topology synthesis as an iterative construction process.<n>At each step, the generation is steered by a lightweight proxy model that predicts multi-objective rewards.<n>Experiments show that GTD can generate highly task-adaptive, sparse, and efficient communication topologies.
arXiv Detail & Related papers (2025-10-09T05:28:28Z) - AMAS: Adaptively Determining Communication Topology for LLM-based Multi-Agent System [19.336020954831202]
Large language models (LLMs) have revolutionized natural language processing capabilities, their practical implementation as autonomous multi-agent systems (MAS) for industrial problem-solving encounters persistent barriers.<n>We introduce AMAS, a paradigm-shifting framework that redefines LLM-based MAS through a novel dynamic graph designer.<n>AMAS exploits the intrinsic properties of individual inputs to intelligently direct query trajectories through task-optimized agent pathways.
arXiv Detail & Related papers (2025-10-02T02:50:22Z) - Assemble Your Crew: Automatic Multi-agent Communication Topology Design via Autoregressive Graph Generation [72.44384066166147]
Multi-agent systems (MAS) based on large language models (LLMs) have emerged as a powerful solution for dealing with complex problems across diverse domains.<n>Existing approaches are fundamentally constrained by their reliance on a template graph modification paradigm with a predefined set of agents and hard-coded interaction structures.<n>We propose ARG-Designer, a novel autoregressive model that operationalizes this paradigm by constructing the collaboration graph from scratch.
arXiv Detail & Related papers (2025-07-24T09:17:41Z) - Efficient Training of Large-Scale AI Models Through Federated Mixture-of-Experts: A System-Level Approach [52.79991638077892]
This article highlights a critical, yet underexplored concept: the absence of robust quantitative strategies for dynamic client-expert alignment.<n>We propose a conceptual system design for intelligent client-expert alignment that incorporates dynamic fitness scoring, global expert load monitoring, and client capacity profiling.
arXiv Detail & Related papers (2025-07-08T05:30:37Z) - NDCG-Consistent Softmax Approximation with Accelerated Convergence [67.10365329542365]
We propose novel loss formulations that align directly with ranking metrics.<n>We integrate the proposed RG losses with the highly efficient Alternating Least Squares (ALS) optimization method.<n> Empirical evaluations on real-world datasets demonstrate that our approach achieves comparable or superior ranking performance.
arXiv Detail & Related papers (2025-06-11T06:59:17Z) - Flow State: Humans Enabling AI Systems to Program Themselves [0.24578723416255752]
We introduce Pocketflow, a platform centered on Human-AI co-design.<n>Pocketflow is a Python framework built upon a deliberately minimal yet synergistic set of core abstractions.<n>It provides a robust, vendor-agnostic foundation with very little code that demonstrably reduces overhead.
arXiv Detail & Related papers (2025-04-03T05:25:46Z) - VERUS-LM: a Versatile Framework for Combining LLMs with Symbolic Reasoning [8.867818326729367]
We introduce VERUS-LM, a novel framework for neurosymbolic reasoning.<n> VERUS-LM employs a generic prompting mechanism, clearly separates domain knowledge from queries.<n>We show that our approach succeeds in diverse reasoning on a novel dataset, markedly outperforming LLMs.
arXiv Detail & Related papers (2025-01-24T14:45:21Z) - Sparse Mixture-of-Experts for Compositional Generalization: Empirical Evidence and Theoretical Foundations of Optimal Sparsity [89.81738321188391]
This study investigates the relationship between task complexity and optimal sparsity in SMoE models.<n>We show that the optimal sparsity lies between minimal activation (1-2 experts) and full activation, with the exact number scaling proportionally to task complexity.
arXiv Detail & Related papers (2024-10-17T18:40:48Z) - AT-MoE: Adaptive Task-planning Mixture of Experts via LoRA Approach [0.6906005491572401]
This paper introduces the Adaptive Task-planing Mixture of Experts(AT-MoE) architecture.
We first train task-specific experts via LoRA approach to enhance problem-solving capabilities and interpretability in specialized areas.
We then introduce a layer-wise adaptive grouped routing module that optimize module fusion based on complex task instructions.
arXiv Detail & Related papers (2024-10-12T13:03:15Z) - Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity [59.57065228857247]
Retrieval-augmented Large Language Models (LLMs) have emerged as a promising approach to enhancing response accuracy in several tasks, such as Question-Answering (QA)
We propose a novel adaptive QA framework, that can dynamically select the most suitable strategy for (retrieval-augmented) LLMs based on the query complexity.
We validate our model on a set of open-domain QA datasets, covering multiple query complexities, and show that ours enhances the overall efficiency and accuracy of QA systems.
arXiv Detail & Related papers (2024-03-21T13:52:30Z) - Outlier-Aware Training for Low-Bit Quantization of Structural
Re-Parameterized Networks [7.446898033580747]
We propose an operator-level improvement for training called Outlier Aware Batch Normalization (OABN)
We also develop a clustering-based non-uniform quantization framework for Quantization-Aware Training (QAT) named ClusterQAT.
arXiv Detail & Related papers (2024-02-11T13:26:40Z) - Inducing Systematicity in Transformers by Attending to Structurally
Quantized Embeddings [60.698130703909804]
Transformers generalize to novel compositions of structures and entities after being trained on a complex dataset.
We propose SQ-Transformer that explicitly encourages systematicity in the embeddings and attention layers.
We show that SQ-Transformer achieves stronger compositional generalization than the vanilla Transformer on multiple low-complexity semantic parsing and machine translation datasets.
arXiv Detail & Related papers (2024-02-09T15:53:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.