Learning to Reason and Memorize with Self-Notes
- URL: http://arxiv.org/abs/2305.00833v2
- Date: Tue, 31 Oct 2023 04:06:28 GMT
- Title: Learning to Reason and Memorize with Self-Notes
- Authors: Jack Lanchantin, Shubham Toshniwal, Jason Weston, Arthur Szlam,
Sainbayar Sukhbaatar
- Abstract summary: Large language models have been shown to struggle with multi-step reasoning.
We propose a simple method for solving both of these problems by allowing the model to take Self-Notes.
- Score: 51.17609489687686
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large language models have been shown to struggle with multi-step reasoning,
and do not retain previous reasoning steps for future use. We propose a simple
method for solving both of these problems by allowing the model to take
Self-Notes. Unlike recent chain-of-thought or scratchpad approaches, the model
can deviate from the input context at any time to explicitly think and write
down its thoughts. This allows the model to perform reasoning on the fly as it
reads the context and even integrate previous reasoning steps, thus enhancing
its memory with useful information and enabling multi-step reasoning.
Experiments across a wide variety of tasks demonstrate that our method can
outperform chain-of-thought and scratchpad methods by taking Self-Notes that
interleave the input text.
Related papers
- Contrastive Chain-of-Thought Prompting [74.10511560147293]
We propose contrastive chain of thought to enhance language model reasoning.
Compared to the conventional chain of thought, our approach provides both valid and invalid reasoning demonstrations.
Our experiments on reasoning benchmarks demonstrate that contrastive chain of thought can serve as a general enhancement of chain-of-thought prompting.
arXiv Detail & Related papers (2023-11-15T18:54:01Z) - Preventing Language Models From Hiding Their Reasoning [0.0]
Large language models (LLMs) often benefit from intermediate steps of reasoning to generate answers to complex problems.
In this work, we focus on one potential way intermediate steps of reasoning could be unfaithful: encoded reasoning.
We show that language models can be trained to make use of encoded reasoning to get higher performance without the user understanding the intermediate steps of reasoning.
arXiv Detail & Related papers (2023-10-27T22:02:29Z) - Large Language Models as Analogical Reasoners [155.9617224350088]
Chain-of-thought (CoT) prompting for language models demonstrates impressive performance across reasoning tasks.
We introduce a new prompting approach, analogical prompting, designed to automatically guide the reasoning process of large language models.
arXiv Detail & Related papers (2023-10-03T00:57:26Z) - Visual Chain of Thought: Bridging Logical Gaps with Multimodal
Infillings [61.04460792203266]
We introduce VCoT, a novel method that leverages chain-of-thought prompting with vision-language grounding to bridge the logical gaps within sequential data.
Our method uses visual guidance to generate synthetic multimodal infillings that add consistent and novel information to reduce the logical gaps for downstream tasks.
arXiv Detail & Related papers (2023-05-03T17:58:29Z) - Chain of Thought Prompting Elicits Reasoning in Large Language Models [56.811278668446825]
This paper explores the ability of language models to generate a coherent chain of thought.
Experiments show that inducing a chain of thought via prompting can enable sufficiently large language models to better perform reasoning tasks.
arXiv Detail & Related papers (2022-01-28T02:33:07Z) - Remembering for the Right Reasons: Explanations Reduce Catastrophic
Forgetting [100.75479161884935]
We propose a novel training paradigm called Remembering for the Right Reasons (RRR)
RRR stores visual model explanations for each example in the buffer and ensures the model has "the right reasons" for its predictions.
We demonstrate how RRR can be easily added to any memory or regularization-based approach and results in reduced forgetting.
arXiv Detail & Related papers (2020-10-04T10:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.