论文标题
将推理能力从具有组成推理变压器的语言模型中解散
Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers
论文作者
论文摘要
本文介绍了推理形式,这是一个统一的推理框架,用于反映复杂决策中人类的模块化和组成推理过程。受认知科学双重过程理论的启发,表示模块(自动思维)和推理模块(受控思维)被解耦以捕获不同水平的认知水平。在表示模块的顶部,预先训练的推理模块在特定和基本的推理技能(例如逻辑,简单质量质量质量等)方面是模块化的,专业的。为了模仿受控的组成思维过程,以平行和级联的方式动态激活和组成不同的推理模块,以控制激活哪些推理技能,以及如何达到推理过程的深度来解决当前问题。统一的推理框架通过单个模型解决了多个任务,并以端到端的方式进行了训练和推断。在11个数据集上进行了评估,需要不同的推理能力和复杂性,推理表现出了实质性的提高,从而揭示了组成推理能力。很少有射击实验通过学习通过有限的数据来构成新任务的预训练技能,并将表示模块和推理模块解耦。进一步的分析显示了推理模块的模块化,因为不同的任务会在不同的推理深度激活不同的推理技能。
This paper presents ReasonFormer, a unified reasoning framework for mirroring the modular and compositional reasoning process of humans in complex decision-making. Inspired by dual-process theory in cognitive science, the representation module (automatic thinking) and reasoning modules (controlled thinking) are decoupled to capture different levels of cognition. Upon the top of the representation module, the pre-trained reasoning modules are modular and professional in specific and fundamental reasoning skills (e.g., logic, simple QA, etc). To mimic the controlled compositional thinking process, different reasoning modules are dynamically activated and composed in both parallel and cascaded manners to control what reasoning skills are activated and how deep the reasoning process will be reached to solve the current problems. The unified reasoning framework solves multiple tasks with a single model, and is trained and inferred in an end-to-end manner. Evaluated on 11 datasets requiring different reasoning skills and complexity, ReasonFormer demonstrates substantial performance boosts, revealing the compositional reasoning ability. Few-shot experiments exhibit better generalization ability by learning to compose pre-trained skills for new tasks with limited data, and decoupling the representation module and the reasoning modules. Further analysis shows the modularity of reasoning modules as different tasks activate distinct reasoning skills at different reasoning depths.