Exchange-of-Thought: Enhancing Large Language Model Capabilities through
Cross-Model Communication
- URL: http://arxiv.org/abs/2312.01823v1
- Date: Mon, 4 Dec 2023 11:53:56 GMT
- Title: Exchange-of-Thought: Enhancing Large Language Model Capabilities through
Cross-Model Communication
- Authors: Zhangyue Yin, Qiushi Sun, Cheng Chang, Qipeng Guo, Junqi Dai, Xuanjing
Huang, Xipeng Qiu
- Abstract summary: Large Language Models (LLMs) have recently made significant strides in complex reasoning tasks through the Chain-of-Thought technique.
We propose Exchange-of-Thought (EoT), a novel framework that enables cross-model communication during problem-solving.
- Score: 76.04373033082948
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large Language Models (LLMs) have recently made significant strides in
complex reasoning tasks through the Chain-of-Thought technique. Despite this
progress, their reasoning is often constrained by their intrinsic
understanding, lacking external insights. To address this, we propose
Exchange-of-Thought (EoT), a novel framework that enables cross-model
communication during problem-solving. Drawing inspiration from network
topology, EoT integrates four unique communication paradigms: Memory, Report,
Relay, and Debate. This paper delves into the communication dynamics and volume
associated with each paradigm. To counterbalance the risks of incorrect
reasoning chains, we implement a robust confidence evaluation mechanism within
these communications. Our experiments across diverse complex reasoning tasks
demonstrate that EoT significantly surpasses established baselines,
underscoring the value of external insights in enhancing LLM performance.
Furthermore, we show that EoT achieves these superior results in a
cost-effective manner, marking a promising advancement for efficient and
collaborative AI problem-solving.
Related papers
- GIVE: Structured Reasoning with Knowledge Graph Inspired Veracity Extrapolation [108.2008975785364]
Graph Inspired Veracity Extrapolation (GIVE) is a novel reasoning framework that integrates the parametric and non-parametric memories.
Our method facilitates a more logical and step-wise reasoning approach akin to experts' problem-solving, rather than gold answer retrieval.
arXiv Detail & Related papers (2024-10-11T03:05:06Z) - The Role of Deductive and Inductive Reasoning in Large Language Models [35.43513487137371]
Large Language Models (LLMs) have achieved substantial progress in artificial intelligence, particularly in reasoning tasks.
We propose the Deductive and InDuctive(DID) method, which enhances LLM reasoning by dynamically integrating both deductive and inductive reasoning.
Our findings suggest that DID provides a more robust and cognitively aligned framework for reasoning in LLMs.
arXiv Detail & Related papers (2024-10-03T18:30:47Z) - Iteration of Thought: Leveraging Inner Dialogue for Autonomous Large Language Model Reasoning [0.0]
Iterative human engagement is a common and effective means of leveraging the advanced language processing power of large language models (LLMs)
We propose the Iteration of Thought (IoT) framework for enhancing LLM responses by generating "thought"-provoking prompts.
Unlike static or semi-static approaches, IoT adapts its reasoning path dynamically, based on evolving context.
arXiv Detail & Related papers (2024-09-19T09:44:17Z) - Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning [79.38140606606126]
We propose an algorithmic framework that fine-tunes vision-language models (VLMs) with reinforcement learning (RL)
Our framework provides a task description and then prompts the VLM to generate chain-of-thought (CoT) reasoning.
We demonstrate that our proposed framework enhances the decision-making capabilities of VLM agents across various tasks.
arXiv Detail & Related papers (2024-05-16T17:50:19Z) - Cantor: Inspiring Multimodal Chain-of-Thought of MLLM [83.6663322930814]
We argue that converging visual context acquisition and logical reasoning is pivotal for tackling visual reasoning tasks.
We propose an innovative multimodal CoT framework, termed Cantor, characterized by a perception-decision architecture.
Our experiments demonstrate the efficacy of the proposed framework, showing significant improvements in multimodal CoT performance.
arXiv Detail & Related papers (2024-04-24T17:59:48Z) - ERA-CoT: Improving Chain-of-Thought through Entity Relationship Analysis [20.24915029448926]
Large language models (LLMs) have achieved commendable accomplishments in various natural language processing tasks.
These challenges arise from the presence of implicit relationships that demand multi-step reasoning.
We propose a novel approach ERA-CoT, which aids LLMs in understanding context by capturing relationships between entities.
arXiv Detail & Related papers (2024-03-11T17:18:53Z) - DDCoT: Duty-Distinct Chain-of-Thought Prompting for Multimodal Reasoning
in Language Models [28.712359821231182]
Large language models (LLMs) have made remarkable strides in such multi-step reasoning on the language modality solely by leveraging the chain of thought (CoT) to mimic human thinking.
The transfer of these advancements to multimodal contexts introduces heightened challenges, including but not limited to the impractical need for labor-intensive annotation.
This study proposes a novel DDCoT prompting that maintains a critical attitude through negative-space prompting and incorporates multimodality into reasoning.
arXiv Detail & Related papers (2023-10-25T08:03:10Z) - Corex: Pushing the Boundaries of Complex Reasoning through Multi-Model Collaboration [83.4031923134958]
Corex is a suite of novel general-purpose strategies that transform Large Language Models into autonomous agents.
Inspired by human behaviors, Corex is constituted by diverse collaboration paradigms including Debate, Review, and Retrieve modes.
We demonstrate that orchestrating multiple LLMs to work in concert yields substantially better performance compared to existing methods.
arXiv Detail & Related papers (2023-09-30T07:11:39Z) - Joint Communication and Computation Framework for Goal-Oriented Semantic
Communication with Distortion Rate Resilience [13.36706909571975]
We use the rate-distortion theory to analyze distortions induced by communication and semantic compression.
We can preemptively estimate the empirical accuracy of AI tasks, making the goal-oriented semantic communication problem feasible.
arXiv Detail & Related papers (2023-09-26T00:26:29Z) - Re-mine, Learn and Reason: Exploring the Cross-modal Semantic
Correlations for Language-guided HOI detection [57.13665112065285]
Human-Object Interaction (HOI) detection is a challenging computer vision task.
We present a framework that enhances HOI detection by incorporating structured text knowledge.
arXiv Detail & Related papers (2023-07-25T14:20:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.