Towards Better Chain-of-Thought Prompting Strategies: A Survey
- URL: http://arxiv.org/abs/2310.04959v1
- Date: Sun, 8 Oct 2023 01:16:55 GMT
- Title: Towards Better Chain-of-Thought Prompting Strategies: A Survey
- Authors: Zihan Yu, Liang He, Zhen Wu, Xinyu Dai, Jiajun Chen
- Abstract summary: Chain-of-Thought (CoT) shows its impressive strength when used as a prompting strategy for large language models (LLM)
Recent years, the prominent effect of CoT prompting has attracted emerging research.
This survey could provide an overall reference on related research.
- Score: 60.75420407216108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Chain-of-Thought (CoT), a step-wise and coherent reasoning chain, shows its
impressive strength when used as a prompting strategy for large language models
(LLM). Recent years, the prominent effect of CoT prompting has attracted
emerging research. However, there still lacks of a systematic summary about key
factors of CoT prompting and comprehensive guide for prompts utilizing. For a
deeper understanding about CoT prompting, we survey on a wide range of current
research, presenting a systematic and comprehensive analysis on several factors
that may influence the effect of CoT prompting, and introduce how to better
apply it in different applications under these discussions. We further analyze
the challenges and propose some future directions about CoT prompting. This
survey could provide an overall reference on related research.
Related papers
- A Hopfieldian View-based Interpretation for Chain-of-Thought Reasoning [48.51969964676017]
Chain-of-Thought (CoT) holds a significant place in augmenting the reasoning performance for large language models.
We propose a Read-and-Control approach for controlling the accuracy of CoT.
arXiv Detail & Related papers (2024-06-18T04:07:13Z) - Towards Faithful Chain-of-Thought: Large Language Models are Bridging Reasoners [19.40385041079461]
Large language models (LLMs) suffer from serious unfaithful chain-of-thought (CoT) issues.
We first study the CoT faithfulness issue at the granularity of CoT steps, identify two reasoning paradigms.
We then conduct a joint analysis of the causal relevance among the context, CoT, and answer during reasoning.
arXiv Detail & Related papers (2024-05-29T09:17:46Z) - Beyond Chain-of-Thought: A Survey of Chain-of-X Paradigms for LLMs [39.214512676276726]
Chain-of-Thought (CoT) has been a widely adopted prompting method, eliciting impressive reasoning abilities of Large Language Models (LLMs)
Inspired by the sequential thought structure of CoT, a number of Chain-of-X (CoX) methods have been developed to address various challenges across diverse domains and tasks involving LLMs.
arXiv Detail & Related papers (2024-04-24T06:12:00Z) - Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models [17.60243337898751]
We present a Chain-of-Action framework for multimodal and retrieval-augmented Question-Answering (QA)
Compared to the literature, CoA overcomes two major challenges of current QA applications: (i) unfaithful hallucination that is inconsistent with real-time or domain facts and (ii) weak reasoning performance over compositional information.
arXiv Detail & Related papers (2024-03-26T03:51:01Z) - ChainLM: Empowering Large Language Models with Improved Chain-of-Thought Prompting [124.69672273754144]
Chain-of-Thought (CoT) prompting can enhance the reasoning capabilities of large language models (LLMs)
Existing CoT approaches usually focus on simpler reasoning tasks and thus result in low-quality and inconsistent CoT prompts.
We introduce CoTGenius, a novel framework designed for the automatic generation of superior CoT prompts.
arXiv Detail & Related papers (2024-03-21T11:34:26Z) - Igniting Language Intelligence: The Hitchhiker's Guide From
Chain-of-Thought Reasoning to Language Agents [80.5213198675411]
Large language models (LLMs) have dramatically enhanced the field of language intelligence.
LLMs leverage the intriguing chain-of-thought (CoT) reasoning techniques, obliging them to formulate intermediate steps en route to deriving an answer.
Recent research endeavors have extended CoT reasoning methodologies to nurture the development of autonomous language agents.
arXiv Detail & Related papers (2023-11-20T14:30:55Z) - Thread of Thought Unraveling Chaotic Contexts [133.24935874034782]
"Thread of Thought" (ThoT) strategy draws inspiration from human cognitive processes.
In experiments, ThoT significantly improves reasoning performance compared to other prompting techniques.
arXiv Detail & Related papers (2023-11-15T06:54:44Z) - Towards Understanding Chain-of-Thought Prompting: An Empirical Study of
What Matters [82.84696222087396]
Chain-of-Thought (CoT) prompting can dramatically improve the multi-step reasoning abilities of large language models (LLMs)
We show that CoT reasoning is possible even with invalid demonstrations.
arXiv Detail & Related papers (2022-12-20T05:20:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.