PARCER as an Operational Contract to Reduce Variance, Cost, and Risk in LLM Systems
- URL: http://arxiv.org/abs/2603.00856v1
- Date: Sun, 01 Mar 2026 01:11:53 GMT
- Title: PARCER as an Operational Contract to Reduce Variance, Cost, and Risk in LLM Systems
- Authors: Elzo Brito dos Santos Filho,
- Abstract summary: This article proposes PARCER as an engineering response to these limitations.<n>The framework acts as a declarative "operational contract" in YAML, transforming interactions into versioned and executable artifacts.<n>The objective of this work is to present the conceptual and technical architecture of PARCER, positioning it as a necessary transition from simple "prompt engineering" to "context engineering with governable governance"
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Systems based on Large Language Models (LLMs) have become formidable tools for automating research and software production. However, their governance remains a challenge when technical requirements demand absolute consistency, auditability, and predictable control over cost and latency. Recent literature highlights two phenomena that aggravate this scenario: the stochastic variance inherent in the model's judgment (often treated as "systemic noise") and the substantial degradation of context utilization in long inputs, with critical losses when decisive information is diluted in the middle of the prompt. This article proposes PARCER as an engineering response to these limitations. The framework acts as a declarative "operational contract" in YAML, transforming unstructured interactions into versioned and executable artifacts. PARCER imposes strict governance structured into seven operational phases, introducing decision hygiene practices inspired by legal judgments to mitigate noise, adaptive token budgeting, formalized recovery routes (fallbacks) for context preservation, and systemic observability via OpenTelemetry. The objective of this work is to present the conceptual and technical architecture of PARCER, positioning it as a necessary transition from simple "prompt engineering" to "context engineering with governable governance".
Related papers
- ReIn: Conversational Error Recovery with Reasoning Inception [43.5498321001366]
This work focuses on error recovery, which necessitates the accurate diagnosis of erroneous dialogue contexts and execution of proper recovery plans.<n>We propose Reasoning Inception (ReIn), a test-time intervention method that plants an initial reasoning into the agent's decision-making process.<n>We evaluate ReIn by systematically simulating conversational failure scenarios that directly hinder successful completion of user goals.
arXiv Detail & Related papers (2026-02-19T02:37:29Z) - Laser: Governing Long-Horizon Agentic Search via Structured Protocol and Context Register [38.329346729947304]
We introduce Laser, a framework for stabilizing and scaling agentic search.<n>Laser organizes agent behaviors into three spaces: planning, task-solving, and retrospection.<n>Laser consistently outperforms existing agentic search baselines under both prompting-only and fine-tuning settings.
arXiv Detail & Related papers (2025-12-23T15:53:33Z) - The Gatekeeper Knows Enough [0.0]
Gatekeeper Protocol is a domain-agnostic framework that governs agent-system interactions.<n>Our results show that this approach significantly increases agent reliability, improves computational efficiency by minimizing token consumption, and enables scalable interaction with complex systems.
arXiv Detail & Related papers (2025-10-16T17:00:42Z) - Technical Report: Facilitating the Adoption of Causal Inference Methods Through LLM-Empowered Co-Pilot [44.336297829718795]
We introduce CATE-B, an open-source co-pilot system that uses large language models (LLMs) within an agentic framework to guide users through treatment effect estimation.<n>CATE-B assists in (i) constructing a structural causal model via causal discovery and LLM-based edge orientation, (ii) identifying robust adjustment sets through a novel Minimal Uncertainty Adjustment Set criterion, and (iii) selecting appropriate regression methods tailored to the causal structure and dataset characteristics.
arXiv Detail & Related papers (2025-08-14T12:20:51Z) - Explainability as a Compliance Requirement: What Regulated Industries Need from AI Tools for Design Artifact Generation [0.7874708385247352]
We investigate the explainability gap in AI-driven design artifact generation through semistructured interviews with ten practitioners from safety-critical industries.<n>Our findings reveal that non-explainable AI outputs necessitate extensive manual validation, reduce stakeholder trust, struggle to handle domain-specific terminology, disrupt team collaboration, and introduce regulatory compliance risks.<n>This study outlines a practical roadmap for improving the transparency, reliability, and applicability of AI tools in requirements engineering.
arXiv Detail & Related papers (2025-07-12T09:34:39Z) - Lightweight Task-Oriented Semantic Communication Empowered by Large-Scale AI Models [66.57755931421285]
Large-scale artificial intelligence (LAI) models pose significant challenges for real-time communication scenarios.<n>This paper proposes utilizing knowledge distillation (KD) techniques to extract and condense knowledge from LAI models.<n>We propose a fast distillation method featuring a pre-stored compression mechanism that eliminates the need for repetitive inference.
arXiv Detail & Related papers (2025-06-16T08:42:16Z) - Hybrid Reasoning for Perception, Explanation, and Autonomous Action in Manufacturing [0.0]
CIPHER is a vision-language-action (VLA) model framework aiming to replicate human-like reasoning for industrial control.<n>It integrates a process expert, a regression model enabling quantitative characterization of system states.<n>It interprets visual or textual inputs from process monitoring, explains its decisions, and autonomously generates precise machine instructions.
arXiv Detail & Related papers (2025-06-10T05:37:33Z) - Evaluating Large Language Models for Real-World Engineering Tasks [75.97299249823972]
This paper introduces a curated database comprising over 100 questions derived from authentic, production-oriented engineering scenarios.<n>Using this dataset, we evaluate four state-of-the-art Large Language Models (LLMs)<n>Our results show that LLMs demonstrate strengths in basic temporal and structural reasoning but struggle significantly with abstract reasoning, formal modeling, and context-sensitive engineering logic.
arXiv Detail & Related papers (2025-05-12T14:05:23Z) - Leveraging LLM Agents and Digital Twins for Fault Handling in Process Plants [0.6990493129893112]
We propose a framework that integrates Large Language Model (LLM) agents with a Digital Twin environment.<n>The Digital Twin acts as a structured repository of plant-specific engineering knowledge for agent prompting.<n>The proposed framework is capable not only of autonomously controlling the mixing module, but also of generating effective corrective actions.
arXiv Detail & Related papers (2025-05-04T12:02:21Z) - Semantic Integrity Constraints: Declarative Guardrails for AI-Augmented Data Processing Systems [39.23499993745249]
We introduce semantic integrity constraints (SICs) for specifying and enforcing correctness conditions over LLM outputs in semantic queries.<n>SICs generalize traditional database integrity constraints to semantic settings, supporting common types of constraints, such as grounding, soundness, and exclusion.<n>We present a system design for integrating SICs into query planning and runtime and discuss its realization in AI-augmented DPSs.
arXiv Detail & Related papers (2025-03-01T19:59:25Z) - ARCEAK: An Automated Rule Checking Framework Enhanced with Architectural Knowledge [2.0159170788984024]
Automated Rule Checking (ARC) plays a crucial role in advancing the construction industry by addressing the laborious, inconsistent, and error-prone nature of traditional model review conducted by industry professionals.<n>Our study introduces a novel approach that decomposes ARC into two distinct tasks: rule information extraction and verification code generation.
arXiv Detail & Related papers (2024-12-10T10:37:11Z) - Defining and executing temporal constraints for evaluating engineering
artifact compliance [56.08728135126139]
Process compliance focuses on ensuring that the actual engineering work is followed as closely as possible to the described engineering processes.
Checking these process constraints is still a daunting task that requires a lot of manual work and delivers feedback to engineers only late in the process.
We present an automated constraint checking approach that can incrementally check temporal constraints across inter-related engineering artifacts upon every artifact change.
arXiv Detail & Related papers (2023-12-20T13:26:31Z) - Counterfactual Explanations as Interventions in Latent Space [62.997667081978825]
Counterfactual explanations aim to provide to end users a set of features that need to be changed in order to achieve a desired outcome.
Current approaches rarely take into account the feasibility of actions needed to achieve the proposed explanations.
We present Counterfactual Explanations as Interventions in Latent Space (CEILS), a methodology to generate counterfactual explanations.
arXiv Detail & Related papers (2021-06-14T20:48:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.