When Efficient Communication Explains Convexity
- URL: http://arxiv.org/abs/2602.02821v1
- Date: Mon, 02 Feb 2026 21:20:45 GMT
- Title: When Efficient Communication Explains Convexity
- Authors: Ashvin Ranjan, Shane Steinert-Threlkeld,
- Abstract summary: The present paper asks what factors are responsible for successful explanations in terms of efficient communication.<n>We first demonstrate and analyze a correlation between optimality in the IB sense and a novel generalization of convexity to this setting.<n>We find that the convexity of the communicative need distribution plays an especially important role.
- Score: 2.1771821757134915
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Much recent work has argued that the variation in the languages of the world can be explained from the perspective of efficient communication; in particular, languages can be seen as optimally balancing competing pressures to be simple and to be informative. Focusing on the expression of meaning -- semantic typology -- the present paper asks what factors are responsible for successful explanations in terms of efficient communication. Using the Information Bottleneck (IB) approach to formalizing this trade-off, we first demonstrate and analyze a correlation between optimality in the IB sense and a novel generalization of convexity to this setting. In a second experiment, we manipulate various modeling parameters in the IB framework to determine which factors drive the correlation between convexity and optimality. We find that the convexity of the communicative need distribution plays an especially important role. These results move beyond showing that efficient communication can explain aspects of semantic typology into explanations for why that is the case by identifying which underlying factors are responsible.
Related papers
- Understanding Usefulness in Developer Explanations on Stack Overflow [2.153604655925499]
This study provides an empirical account of what drives perceived usefulness in developer explanations.<n>It offers implications for how developers and RE practitioners can craft clearer and more effective explanations.
arXiv Detail & Related papers (2026-01-21T10:50:43Z) - Implicature in Interaction: Understanding Implicature Improves Alignment in Human-LLM Interaction [8.735329612895578]
Implicature (meaning conveyed beyond explicit statements through shared context) is essential for human-AI alignment.<n>This study examines Large Language Models' ability to infer user intent embedded in context-driven prompts.<n>Results show that larger models approximate human interpretations more closely, while smaller models struggle with implicature inference.
arXiv Detail & Related papers (2025-10-29T11:49:42Z) - Understanding the Information Propagation Effects of Communication Topologies in LLM-based Multi-Agent Systems [58.95962217043371]
We present a causal framework to analyze how agent outputs, whether correct or erroneous, propagate under topologies with varying sparsity.<n>Our empirical studies reveal that moderately sparse topologies, which effectively suppress error propagation while preserving beneficial information diffusion, typically achieve optimal task performance.<n>We propose a novel topology design approach, EIB-leanrner, that balances error suppression and beneficial information propagation by fusing connectivity patterns from both dense and sparse graphs.
arXiv Detail & Related papers (2025-05-29T11:21:48Z) - A primer on optimal transport for causal inference with observational data [0.0]
The goal of this review is to offer an introduction to the surprisingly deep existing connections between optimal transport and the identification of causal effects with observational data.<n>As a result, this review is intended to unify the language and notation between different areas of statistics, mathematics, and econometrics.
arXiv Detail & Related papers (2025-03-10T19:51:37Z) - Causal Reasoning in Large Language Models: A Knowledge Graph Approach [6.5344638992876085]
Large language models (LLMs) typically improve performance by either retrieving semantically similar information, or enhancing reasoning abilities through structured prompts like chain-of-thought.
This paper proposes a knowledge graph (KG)-based random-walk reasoning approach that leverages causal relationships.
arXiv Detail & Related papers (2024-10-15T13:24:44Z) - Generative causal testing to bridge data-driven models and scientific theories in language neuroscience [82.995061475971]
We present generative causal testing (GCT), a framework for generating concise explanations of language selectivity in the brain.<n>We show that GCT can dissect fine-grained differences between brain areas with similar functional selectivity.
arXiv Detail & Related papers (2024-10-01T15:57:48Z) - PACE: A Pragmatic Agent for Enhancing Communication Efficiency Using
Large Language Models [29.016842120305892]
This paper proposes an image pragmatic communication framework based on a Pragmatic Agent for Communication Efficiency (PACE) using Large Language Models (LLM)
PACE sequentially performs semantic perception, intention resolution, and intention-oriented coding.
For experimental validation, this paper constructs an image pragmatic communication dataset along with corresponding evaluation standards.
arXiv Detail & Related papers (2024-01-30T06:55:17Z) - From Heuristic to Analytic: Cognitively Motivated Strategies for
Coherent Physical Commonsense Reasoning [66.98861219674039]
Heuristic-Analytic Reasoning (HAR) strategies drastically improve the coherence of rationalizations for model decisions.
Our findings suggest that human-like reasoning strategies can effectively improve the coherence and reliability of PLM reasoning.
arXiv Detail & Related papers (2023-10-24T19:46:04Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - Modeling Hierarchical Reasoning Chains by Linking Discourse Units and
Key Phrases for Reading Comprehension [80.99865844249106]
We propose a holistic graph network (HGN) which deals with context at both discourse level and word level, as the basis for logical reasoning.
Specifically, node-level and type-level relations, which can be interpreted as bridges in the reasoning process, are modeled by a hierarchical interaction mechanism.
arXiv Detail & Related papers (2023-06-21T07:34:27Z) - Complementary Explanations for Effective In-Context Learning [77.83124315634386]
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts.
This work aims to better understand the mechanisms by which explanations are used for in-context learning.
arXiv Detail & Related papers (2022-11-25T04:40:47Z) - Counterfactual Reasoning for Out-of-distribution Multimodal Sentiment
Analysis [56.84237932819403]
This paper aims to estimate and mitigate the bad effect of textual modality for strong OOD generalization.
Inspired by this, we devise a model-agnostic counterfactual framework for multimodal sentiment analysis.
arXiv Detail & Related papers (2022-07-24T03:57:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.