Reasoning aligns language models to human cognition
- URL: http://arxiv.org/abs/2602.08693v1
- Date: Mon, 09 Feb 2026 14:13:39 GMT
- Title: Reasoning aligns language models to human cognition
- Authors: Gonçalo Guiomar, Elia Torre, Pehuen Moure, Victoria Shavina, Mario Giulianelli, Shih-Chii Liu, Valerio Mante,
- Abstract summary: We introduce an active probabilistic reasoning task that cleanly separates sampling (actively acquiring evidence) from inference (integrating evidence toward a decision)<n> Benchmarking humans and a broad set of contemporary large language models against near-optimal reference policies reveals a consistent pattern.<n>This model places humans and models in a shared low-dimensional cognitive space, reproduces behavioral signatures across agents, and shows how chain-of-thought shifts language models toward human-like regimes of evidence accumulation and belief-to-choice mapping.
- Score: 12.07126784684808
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Do language models make decisions under uncertainty like humans do, and what role does chain-of-thought (CoT) reasoning play in the underlying decision process? We introduce an active probabilistic reasoning task that cleanly separates sampling (actively acquiring evidence) from inference (integrating evidence toward a decision). Benchmarking humans and a broad set of contemporary large language models against near-optimal reference policies reveals a consistent pattern: extended reasoning is the key determinant of strong performance, driving large gains in inference and producing belief trajectories that become strikingly human-like, while yielding only modest improvements in active sampling. To explain these differences, we fit a mechanistic model that captures systematic deviations from optimal behavior via four interpretable latent variables: memory, strategy, choice bias, and occlusion awareness. This model places humans and models in a shared low-dimensional cognitive space, reproduces behavioral signatures across agents, and shows how chain-of-thought shifts language models toward human-like regimes of evidence accumulation and belief-to-choice mapping, tightening alignment in inference while leaving a persistent gap in information acquisition.
Related papers
- Reasoning as State Transition: A Representational Analysis of Reasoning Evolution in Large Language Models [50.39102836928242]
We introduce a representational perspective to investigate the dynamics of the model's internal states.<n>We discover that post-training yields only limited improvement in static initial representation quality.
arXiv Detail & Related papers (2026-01-31T15:23:33Z) - Reasoning Models Generate Societies of Thought [9.112083442162671]
We show that enhanced reasoning emerges from simulating multi-agent-like interactions.<n>We find that reasoning models like DeepSeek-R1 and QwQ-32B exhibit much greater perspective diversity than instruction-tuned models.
arXiv Detail & Related papers (2026-01-15T19:52:33Z) - Analyzing Reasoning Consistency in Large Multimodal Models under Cross-Modal Conflicts [74.47786985522762]
We identify a critical failure mode termed textual inertia, where models tend to blindly adhere to the erroneous text while neglecting conflicting visual evidence.<n>We propose the LogicGraph Perturbation Protocol that structurally injects perturbations into the reasoning chains of diverse LMMs.<n>Results reveal that models successfully self-correct in less than 10% of cases and predominantly succumb to blind textual error propagation.
arXiv Detail & Related papers (2026-01-07T16:39:34Z) - Forgetting as a Feature: Cognitive Alignment of Large Language Models [39.146761527401424]
We show that Large Language Models (LLMs) exhibit systematic forgetting of past information.<n> Drawing inspiration from human memory dynamics, we model LLM inference as a probabilistic memory process governed by exponential decay.<n>Building on these observations, we propose probabilistic memory prompting, a lightweight strategy that shapes evidence integration to mimic human-like memory decay.
arXiv Detail & Related papers (2025-12-28T10:43:00Z) - LLMs as Strategic Agents: Beliefs, Best Response Behavior, and Emergent Heuristics [0.0]
Large Language Models (LLMs) are increasingly applied to domains that require reasoning about other agents' behavior.<n>We show that current frontier models exhibit belief-coherent best-response behavior at targeted reasoning memorization.<n>Under increasing complexity, explicit recursion gives way to internally generated rules of choice that are stable, model-specific, and distinct from known human biases.
arXiv Detail & Related papers (2025-10-12T21:40:29Z) - Variational Reasoning for Language Models [93.08197299751197]
We introduce a variational reasoning framework for language models that treats thinking traces as latent variables.<n>We show that rejection sampling finetuning and binary-reward RL, including GRPO, can be interpreted as local forward-KL objectives.
arXiv Detail & Related papers (2025-09-26T17:58:10Z) - Extended Inductive Reasoning for Personalized Preference Inference from Behavioral Signals [45.019257216564036]
This paper investigates extended inductive reasoning in large language models (LLMs)<n>We propose AlignXplore, a model that enables systematic preference inference from behavioral signals in users' interaction histories.<n>We show that AlignXplore achieves substantial improvements over the backbone model by an average of 15.49% on in-domain and out-of-domain benchmarks.
arXiv Detail & Related papers (2025-05-23T16:16:46Z) - Internal Causal Mechanisms Robustly Predict Language Model Out-of-Distribution Behaviors [61.92704516732144]
We show that the most robust features for correctness prediction are those that play a distinctive causal role in the model's behavior.<n>We propose two methods that leverage causal mechanisms to predict the correctness of model outputs.
arXiv Detail & Related papers (2025-05-17T00:31:39Z) - Using Reinforcement Learning to Train Large Language Models to Explain Human Decisions [11.40240971657506]
In this work, we explore the potential of pretrained large language models to serve as dual-purpose cognitive models.<n>We employ reinforcement learning with outcome-based rewards to guide LLMs toward generating explicit reasoning traces for explaining human risky choices.
arXiv Detail & Related papers (2025-05-16T18:22:05Z) - Reasoning Capabilities of Large Language Models on Dynamic Tasks [0.017476232824732776]
Large language models excel on static benchmarks, but their ability as self-learning agents in dynamic environments remains unclear.<n>We evaluate three prompting strategies: self-reflection, mutation, and planning across dynamic tasks with open-source models.<n>We find that larger models generally outperform smaller ones, but that strategic prompting can close this performance gap.
arXiv Detail & Related papers (2025-05-15T17:53:47Z) - Conceptual and Unbiased Reasoning in Language Models [98.90677711523645]
We propose a novel conceptualization framework that forces models to perform conceptual reasoning on abstract questions.
We show that existing large language models fall short on conceptual reasoning, dropping 9% to 28% on various benchmarks.
We then discuss how models can improve since high-level abstract reasoning is key to unbiased and generalizable decision-making.
arXiv Detail & Related papers (2024-03-30T00:53:53Z) - Interpreting Language Models with Contrastive Explanations [99.7035899290924]
Language models must consider various features to predict a token, such as its part of speech, number, tense, or semantics.
Existing explanation methods conflate evidence for all these features into a single explanation, which is less interpretable for human understanding.
We show that contrastive explanations are quantifiably better than non-contrastive explanations in verifying major grammatical phenomena.
arXiv Detail & Related papers (2022-02-21T18:32:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.