论文标题

来自不受信任的老师的诚实学生:从验证的语言模型中学习一条可解释的提问管道

Honest Students from Untrusted Teachers: Learning an Interpretable Question-Answering Pipeline from a Pretrained Language Model

论文作者

Eisenstein, Jacob, Andor, Daniel, Bohnet, Bernd, Collins, Michael, Mimno, David

论文摘要

可解释的问题答案系统不仅应产生准确的答案,还应产生理由,以证明其推理并允许人类检查其工作。但是,哪种理由是有用的,我们如何培训系统来生产它们呢?我们提出了一种新的原理,用于打开书本问题,称为\ emph {markup and mask},它结合了提取性和自由文本解释的各个方面。在标记阶段,通过自由文本标记增强了段落,使每个句子都可以在话语上下文之外站立。在掩蔽阶段,选择了标记的段落的子跨度。为了训练一个系统以产生标记和掩盖理由而没有注释,我们利用了文化学习。具体而言,我们通过向验证的验证的语言模型发送一系列提示来生成银注释的数据,该模型充当老师。然后,我们通过对导致正确答案的理由子集进行培训来微调较小的学生模型。从某种意义上说,学生是“诚实的”,因为它是一条管道:基本原理是段落和答案之间的瓶颈,而“不受信任的”老师在没有这样的限制下运作。因此,我们提供了一种新的方式来构建可信赖的管道系统,从终点注释和冻结的语言模型的结合使用。

Explainable question answering systems should produce not only accurate answers but also rationales that justify their reasoning and allow humans to check their work. But what sorts of rationales are useful and how can we train systems to produce them? We propose a new style of rationale for open-book question answering, called \emph{markup-and-mask}, which combines aspects of extractive and free-text explanations. In the markup phase, the passage is augmented with free-text markup that enables each sentence to stand on its own outside the discourse context. In the masking phase, a sub-span of the marked-up passage is selected. To train a system to produce markup-and-mask rationales without annotations, we leverage in-context learning. Specifically, we generate silver annotated data by sending a series of prompts to a frozen pretrained language model, which acts as a teacher. We then fine-tune a smaller student model by training on the subset of rationales that led to correct answers. The student is "honest" in the sense that it is a pipeline: the rationale acts as a bottleneck between the passage and the answer, while the "untrusted" teacher operates under no such constraints. Thus, we offer a new way to build trustworthy pipeline systems from a combination of end-task annotations and frozen pretrained language models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源