论文标题
逻辑推理,具有可解释和强大的NLI模型的跨度级预测
Logical Reasoning with Span-Level Predictions for Interpretable and Robust NLI Models
论文作者
论文摘要
当前的自然语言推断(NLI)模型取得了令人印象深刻的结果,有时在评估分配测试集时表现出色。但是,众所周知,这些模型可以从注释人工制品和数据集偏见中学习,因此尚不清楚模型在多大程度上学习NLI的任务,而不是在其培训数据中从浅启发式学中学习。我们通过引入NLI的逻辑推理框架来解决此问题,从而创建基于逻辑规则的高度透明的模型决策。与先前的工作不同,我们表明可以提高可解释性,而无需降低预测精度。我们几乎完全保留了SNLI的性能,同时还确定了负责每个模型预测的确切假设跨度。使用E-SNLI人类的解释,我们验证我们的模型在训练过程中没有使用任何跨度标签,但我们的模型在跨度水平上做出了明智的决策。我们可以通过在训练过程中使用E-SNLI解释来进一步提高模型性能和跨度级别的决策。最后,我们的模型在减少的数据设置中更强大。当只有1,000个示例训练时,相对于基线,MNLI匹配和不匹配的验证集的分布性能提高了13%和16%。较少的观察训练可以进一步改善分布和分发。
Current Natural Language Inference (NLI) models achieve impressive results, sometimes outperforming humans when evaluating on in-distribution test sets. However, as these models are known to learn from annotation artefacts and dataset biases, it is unclear to what extent the models are learning the task of NLI instead of learning from shallow heuristics in their training data. We address this issue by introducing a logical reasoning framework for NLI, creating highly transparent model decisions that are based on logical rules. Unlike prior work, we show that improved interpretability can be achieved without decreasing the predictive accuracy. We almost fully retain performance on SNLI, while also identifying the exact hypothesis spans that are responsible for each model prediction. Using the e-SNLI human explanations, we verify that our model makes sensible decisions at a span level, despite not using any span labels during training. We can further improve model performance and span-level decisions by using the e-SNLI explanations during training. Finally, our model is more robust in a reduced data setting. When training with only 1,000 examples, out-of-distribution performance improves on the MNLI matched and mismatched validation sets by 13% and 16% relative to the baseline. Training with fewer observations yields further improvements, both in-distribution and out-of-distribution.