论文标题
语言表示的核心推理学习
Coreferential Reasoning Learning for Language Representation
论文作者
论文摘要
诸如BERT之类的语言表示模型可以有效地从纯文本中捕获上下文语义信息,并已被证明可以在许多下游NLP任务中实现有希望的结果,并进行适当的微调。但是,大多数现有的语言表示模型无法明确处理核心,这对于对整个话语的连贯理解至关重要。为了解决这个问题,我们提出了Corefbert,这是一种新型语言表示模型,可以捕获上下文中的核心关系。实验结果表明,与现有的基线模型相比,Corefbert可以在需要核心推理的各种下游NLP任务上始终如一地实现重大改进,同时在其他常见的NLP任务上保持了与以前的模型可比性的可比性。本文的源代码和实验详细信息可以从https://github.com/thunlp/corefbert获得。
Language representation models such as BERT could effectively capture contextual semantic information from plain text, and have been proved to achieve promising results in lots of downstream NLP tasks with appropriate fine-tuning. However, most existing language representation models cannot explicitly handle coreference, which is essential to the coherent understanding of the whole discourse. To address this issue, we present CorefBERT, a novel language representation model that can capture the coreferential relations in context. The experimental results show that, compared with existing baseline models, CorefBERT can achieve significant improvements consistently on various downstream NLP tasks that require coreferential reasoning, while maintaining comparable performance to previous models on other common NLP tasks. The source code and experiment details of this paper can be obtained from https://github.com/thunlp/CorefBERT.