论文标题
带有评论的教学
Teaching with Commentaries
论文作者
论文摘要
对深度神经网络的有效培训可能具有挑战性,并且关于如何最好地学习这些模型还有许多开放的问题。最近开发了改善神经网络培训检查教学的方法:在培训过程中提供学习的信息以改善下游模型的性能。在本文中,我们采取步骤来扩展教学范围。我们建议使用评论的灵活的教学框架,学习的元信息有助于对特定任务进行培训。我们提出了基于梯度的方法来学习评论,并利用有关可扩展性的隐式差异化的最新工作。我们探讨了评论的各种应用,从加权培训示例到参数依赖标签的数据增强策略,再到表示突出显着图像区域的注意力掩码。我们发现评论可以提高训练速度和/或性能,并提供有关数据集和培训过程的见解。我们还观察到评论概括了:在训练新模型以获得绩效益处时,可以重复使用它们,这表明用用例存储在数据集中,并在将来杠杆上存储评论以进行改进的模型培训。
Effective training of deep neural networks can be challenging, and there remain many open questions on how to best learn these models. Recently developed methods to improve neural network training examine teaching: providing learned information during the training process to improve downstream model performance. In this paper, we take steps towards extending the scope of teaching. We propose a flexible teaching framework using commentaries, learned meta-information helpful for training on a particular task. We present gradient-based methods to learn commentaries, leveraging recent work on implicit differentiation for scalability. We explore diverse applications of commentaries, from weighting training examples, to parameterising label-dependent data augmentation policies, to representing attention masks that highlight salient image regions. We find that commentaries can improve training speed and/or performance, and provide insights about the dataset and training process. We also observe that commentaries generalise: they can be reused when training new models to obtain performance benefits, suggesting a use-case where commentaries are stored with a dataset and leveraged in future for improved model training.