论文标题
CKA作为深度学习的相似性措施的可靠性
Reliability of CKA as a Similarity Measure in Deep Learning
论文作者
论文摘要
比较神经网络中学到的神经表示是一个具有挑战性但重要的问题,这已经以不同的方式解决了。中心内核对齐(CKA)相似性度量,尤其是其线性变体,最近已成为一种流行的方法,已被广泛用于比较网络的不同层的表示形式,对经过不同训练的架构相似的网络或与在同一数据上训练的不同建筑的模型。关于这些各种表示的相似性和差异的各种结论已使用CKA得出。在这项工作中,我们提出了正式表征CKA对大量简单转换的敏感性的分析,这些转换自然可以在现代机器学习的背景下发生。这为CKA对离群值的敏感性提供了具体的解释,这在过去的工作中已经观察到,以及保留数据线性可分离性的转换,这是一个重要的概括属性。我们从经验上研究了CKA相似性度量的几个弱点,证明了它给出意外或违反直觉结果的情况。最后,我们研究了修改表示形式以维持功能行为的方法,同时更改CKA值。我们的结果表明,在许多情况下,CKA值可以很容易地操纵,而无需对模型的功能行为进行实质性更改,并在利用激活对准指标时谨慎行事。
Comparing learned neural representations in neural networks is a challenging but important problem, which has been approached in different ways. The Centered Kernel Alignment (CKA) similarity metric, particularly its linear variant, has recently become a popular approach and has been widely used to compare representations of a network's different layers, of architecturally similar networks trained differently, or of models with different architectures trained on the same data. A wide variety of conclusions about similarity and dissimilarity of these various representations have been made using CKA. In this work we present analysis that formally characterizes CKA sensitivity to a large class of simple transformations, which can naturally occur in the context of modern machine learning. This provides a concrete explanation of CKA sensitivity to outliers, which has been observed in past works, and to transformations that preserve the linear separability of the data, an important generalization attribute. We empirically investigate several weaknesses of the CKA similarity metric, demonstrating situations in which it gives unexpected or counter-intuitive results. Finally we study approaches for modifying representations to maintain functional behaviour while changing the CKA value. Our results illustrate that, in many cases, the CKA value can be easily manipulated without substantial changes to the functional behaviour of the models, and call for caution when leveraging activation alignment metrics.