Natural Language Edge Labelling: Decoupling Intent from Execution in Structured LM Reasoning
- URL: http://arxiv.org/abs/2510.04817v1
- Date: Mon, 06 Oct 2025 14:00:02 GMT
- Title: Natural Language Edge Labelling: Decoupling Intent from Execution in Structured LM Reasoning
- Authors: Abhinav Madahar,
- Abstract summary: We introduce Natural Language Edge Labelling (NLEL), a labeller-tuner overlay that attaches a free-form natural-language directive to each search edge.<n>We show NLEL strictly generalizes CoT/ToT, prove an anytime-monotonicity property for top-$k$ selection under labelconditioned bundles, and bound selector shortfall by controlvector distortion.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Controllers for structured LM reasoning (e.g., Chain-of-Thought, self-consistency, and Tree-of-Thoughts) often entangle what to try next with how to execute it, exposing only coarse global knobs and yielding brittle, compute-inefficient, and hard-to-audit behavior. We introduce Natural Language Edge Labelling (NLEL), a labeller-tuner overlay that attaches a free-form natural-language directive to each search edge and translates it into a schema-bounded control vector for decoding, search (branch quotas, exploration $\beta$), generation bundle size, retrieval mixtures, and verification passes. A labeller $\Lambda$ emits labels from the parent state and a compact context; a tuner $\Psi$ maps $(P, L, C)\to \Pi$, with strict schema validation and trust-region projection around safe defaults. Downstream selection remains ToT-style with score $S=\mu+\beta\sigma$ and depth-annealed $\beta$. We show NLEL strictly generalizes CoT/ToT, prove an anytime-monotonicity property for top-$k$ selection under label-conditioned bundles, and bound selector shortfall by control-vector distortion, providing decision-relevant justification for guards like trust regions and verification passes. We instantiate $\Psi$ as a prompt-only JSON Parameter Emitter and preregister an evaluation on GSM8K, MATH (subset), StrategyQA, and ARC-Challenge with compute-aware reporting (success@compute, tokens-per-success) and ablations over $\Lambda$, $\Psi$, trust-region radius, and control quantization; preregistered forecasts anticipate accuracy gains at comparable token budgets and improved success@compute under constraints. NLEL offers an interpretable, model-agnostic interface that separates intent from execution for controllable, auditable LM inference.
Related papers
- Learning Accurate Segmentation Purely from Self-Supervision [87.78965637247107]
Selfment is a fully self-supervised framework that segments objects directly from raw images without human labels.<n>Selfment sets new state-of-the-art (SoTA) results across multiple benchmarks.
arXiv Detail & Related papers (2026-02-27T07:36:32Z) - TabTracer: Monte Carlo Tree Search for Complex Table Reasoning with Large Language Models [10.584052101655537]
TabTracer is an agentic framework that coordinates multi-step tool calls over intermediate table states.<n>It enforces step-level verification with typed operations and lightweight numeric and format checks.<n>It reduces redundancy with budget-aware pruning, deduplication, and state hashing with a monotonicity gate to cut token cost.
arXiv Detail & Related papers (2026-02-15T10:39:43Z) - Optimal Budgeted Adaptation of Large Language Models [8.207283283207138]
Trade-off between labeled data availability and downstream accuracy remains a central challenge in large language models.<n>We propose a principled framework for emphbudget-aware supervised fine-tuning by casting LLM adaptation as a contextual Stackelberg game.
arXiv Detail & Related papers (2026-02-01T01:08:15Z) - SCOPE: Saliency-Coverage Oriented Token Pruning for Efficient Multimodel LLMs [59.415473779171315]
We propose a novel visual token pruning strategy called textbfSaliency-textbfCoverage textbfOriented token textbfPruning for textbfEfficient MLLMs.
arXiv Detail & Related papers (2025-10-28T09:29:37Z) - Unsupervised Conformal Inference: Bootstrapping and Alignment to Control LLM Uncertainty [49.19257648205146]
We propose an unsupervised conformal inference framework for generation.<n>Our gates achieve close-to-nominal coverage and provide tighter, more stable thresholds than split UCP.<n>The result is a label-free, API-compatible gate for test-time filtering.
arXiv Detail & Related papers (2025-09-26T23:40:47Z) - Auditable Early Stopping for Agentic Routing: Ledger-Verified Run-Wise Certificates under Local DP [0.0]
We address when a best-first router for tool-use agents can stop exploring without missing a better leaf.<n>We introduce a run-wise certificate that couples each node's key to the same exponential race that realizes leaf perturbations.<n>Experiments on synthetic graphs and a small real pipeline show tight stopping, deterministic replay, and low overhead.
arXiv Detail & Related papers (2025-09-09T01:25:09Z) - Fast Controlled Generation from Language Models with Adaptive Weighted Rejection Sampling [90.86991492288487]
evaluating constraint on every token can be prohibitively expensive.<n> LCD can distort the global distribution over strings, sampling tokens based only on local information.<n>We show that our approach is superior to state-of-the-art baselines.
arXiv Detail & Related papers (2025-04-07T18:30:18Z) - $\texttt{SEM-CTRL}$: Semantically Controlled Decoding [53.86639808659575]
$texttSEM-CTRL$ is a unified approach that enforces rich context-sensitive constraints and task- and instance-specific semantics directly on an LLM decoder.<n>texttSEM-CTRL$ allows small pre-trained LLMs to efficiently outperform larger variants and state-of-the-art reasoning models.
arXiv Detail & Related papers (2025-03-03T18:33:46Z) - Reasoning to Attend: Try to Understand How <SEG> Token Works [44.33848900059659]
We show that the $texttSEG>$ token contributes to semantic similarity within image-text pairs.<n>We present READ, which facilitates LMMs' resilient $textbfREA$soning capability of where to atten$textbfD$ under the guidance of highly activated points.
arXiv Detail & Related papers (2024-12-23T17:44:05Z) - Selective Generation for Controllable Language Models [19.909671258499184]
Trustworthiness of generative language models (GLMs) is crucial in their deployment to critical decision making systems.<n>We propose two selective generation algorithms which control the false discovery rate with respect to the textual entailment relation (FDR-E)<n>$textttSGentextttSup$, a direct modification of the selective prediction, exploits entailment-labeled data, annotated by humans.<n>Since human annotation is costly, we propose a semi-supervised version, $textttSGentextttSemi$, which fully utilizes the un
arXiv Detail & Related papers (2023-07-18T13:36:24Z) - Semi-DETR: Semi-Supervised Object Detection with Detection Transformers [105.45018934087076]
We analyze the DETR-based framework on semi-supervised object detection (SSOD)
We present Semi-DETR, the first transformer-based end-to-end semi-supervised object detector.
Our method outperforms all state-of-the-art methods by clear margins.
arXiv Detail & Related papers (2023-07-16T16:32:14Z) - Stochastic Bandits with Linear Constraints [69.757694218456]
We study a constrained contextual linear bandit setting, where the goal of the agent is to produce a sequence of policies.
We propose an upper-confidence bound algorithm for this problem, called optimistic pessimistic linear bandit (OPLB)
arXiv Detail & Related papers (2020-06-17T22:32:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.