Disentangling Reasoning Capabilities from Language Models with
Compositional Reasoning Transformers
- URL: http://arxiv.org/abs/2210.11265v1
- Date: Thu, 20 Oct 2022 13:39:55 GMT
- Title: Disentangling Reasoning Capabilities from Language Models with
Compositional Reasoning Transformers
- Authors: Wanjun Zhong, Tingting Ma, Jiahai Wang, Jian Yin, Tiejun Zhao,
Chin-Yew Lin and Nan Duan
- Abstract summary: ReasonFormer is a unified reasoning framework for mirroring the modular and compositional reasoning process of humans.
The representation module (automatic thinking) and reasoning modules (controlled thinking) are disentangled to capture different levels of cognition.
The unified reasoning framework solves multiple tasks with a single model,and is trained and inferred in an end-to-end manner.
- Score: 72.04044221898059
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents ReasonFormer, a unified reasoning framework for mirroring
the modular and compositional reasoning process of humans in complex decision
making. Inspired by dual-process theory in cognitive science, the
representation module (automatic thinking) and reasoning modules (controlled
thinking) are disentangled to capture different levels of cognition. Upon the
top of the representation module, the pre-trained reasoning modules are modular
and expertise in specific and fundamental reasoning skills (e.g., logic, simple
QA, etc). To mimic the controlled compositional thinking process, different
reasoning modules are dynamically activated and composed in both parallel and
cascaded manners to control what reasoning skills are activated and how deep
the reasoning process will be reached to solve the current problems. The
unified reasoning framework solves multiple tasks with a single model,and is
trained and inferred in an end-to-end manner. Evaluated on 11 datasets
requiring different reasoning skills and complexity, ReasonFormer demonstrates
substantial performance boosts, revealing the compositional reasoning ability.
Few-shot experiments exhibit better generalization ability by learning to
compose pre-trained skills for new tasks with limited data,and decoupling the
representation module and the reasoning modules. Further analysis shows the
modularity of reasoning modules as different tasks activate distinct reasoning
skills at different reasoning depths.
Related papers
- Boosting Deductive Reasoning with Step Signals In RLHF [15.441793744822457]
We have developed an automated method, Multi-step Deduction (MuseD), for deductive reasoning data.
MuseD has allowed us to create training and testing datasets for multi-step reasoning.
Our training data has demonstrated significant improvements in logical capabilities for both in-domain of out-of-domain reasoning tasks.
arXiv Detail & Related papers (2024-10-12T13:19:11Z) - How Truncating Weights Improves Reasoning in Language Models [49.80959223722325]
We study how certain global associations tend to be stored in specific weight components or Transformer blocks.
We analyze how this arises during training, both empirically and theoretically.
arXiv Detail & Related papers (2024-06-05T08:51:08Z) - DDCoT: Duty-Distinct Chain-of-Thought Prompting for Multimodal Reasoning
in Language Models [28.712359821231182]
Large language models (LLMs) have made remarkable strides in such multi-step reasoning on the language modality solely by leveraging the chain of thought (CoT) to mimic human thinking.
The transfer of these advancements to multimodal contexts introduces heightened challenges, including but not limited to the impractical need for labor-intensive annotation.
This study proposes a novel DDCoT prompting that maintains a critical attitude through negative-space prompting and incorporates multimodality into reasoning.
arXiv Detail & Related papers (2023-10-25T08:03:10Z) - Dynamic MOdularized Reasoning for Compositional Structured Explanation
Generation [29.16040150962427]
We propose a dynamic modularized reasoning model, MORSE, to improve compositional generalization of neural models.
MORSE factorizes the inference process into a combination of modules, where each module represents a functional unit.
We conduct experiments for increasing lengths and shapes of reasoning trees on two benchmarks to test MORSE's compositional generalization abilities.
arXiv Detail & Related papers (2023-09-14T11:40:30Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - MetaLogic: Logical Reasoning Explanations with Fine-Grained Structure [129.8481568648651]
We propose a benchmark to investigate models' logical reasoning capabilities in complex real-life scenarios.
Based on the multi-hop chain of reasoning, the explanation form includes three main components.
We evaluate the current best models' performance on this new explanation form.
arXiv Detail & Related papers (2022-10-22T16:01:13Z) - Teaching Neural Module Networks to Do Arithmetic [54.06832128723388]
We up-grade NMNs by bridging the gap between its interpreter and the complex questions.
We introduce addition and subtraction modules that perform numerical reasoning over numbers.
On a subset of DROP, experimental results show that our proposed methods enhance NMNs' numerical reasoning skills by 17.7% improvement of F1 score.
arXiv Detail & Related papers (2022-10-06T06:38:04Z) - Prediction or Comparison: Toward Interpretable Qualitative Reasoning [16.02199526395448]
Current approaches use either semantics to transform natural language inputs into logical expressions or a "black-box" model to solve them in one step.
In this work, we categorize qualitative reasoning tasks into two types: prediction and comparison.
In particular, we adopt neural network modules trained in an end-to-end manner to simulate the two reasoning processes.
arXiv Detail & Related papers (2021-06-04T10:27:55Z) - Towards Interpretable Reasoning over Paragraph Effects in Situation [126.65672196760345]
We focus on the task of reasoning over paragraph effects in situation, which requires a model to understand the cause and effect.
We propose a sequential approach for this task which explicitly models each step of the reasoning process with neural network modules.
In particular, five reasoning modules are designed and learned in an end-to-end manner, which leads to a more interpretable model.
arXiv Detail & Related papers (2020-10-03T04:03:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.