论文标题

与卑鄙的老师进行半监督学习的本地聚类

Local Clustering with Mean Teacher for Semi-supervised Learning

论文作者

Chen, Zexi, Dutton, Benjamin, Ramachandra, Bharathkumar, Wu, Tianfu, Vatsavai, Ranga Raju

论文摘要

Tarvainen和Valpola的平均教师(MT)模型在几个半监督基准数据集上表现出了有利的表现。 MT保持教师模型的权重,因为学生模型的权重的指数移动平均值,并最大程度地减少了在各种输入扰动下的概率预测之间的差异。但是,众所周知,MT遭受确认偏见的困扰,即加强了不正确的教师模型预测。在这项工作中,我们提出了一种称为局部聚类(LC)的简单而有效的方法,以减轻确认偏差的影响。在MT中,每个数据点在训练过程中被视为独立于其他点。但是,如果数据点共享相似的功能,则可能在功能空间中彼此接近。在此激励的情况下,我们通过最大程度地减少特征空间中相邻数据点之间的成对距离来局部群集数据点。结合标记的数据点上的标准分类跨凝结目标,借助邻居的帮助,将错误分类的未标记数据点伸向其正确类别的高密度区域,从而改善了模型性能。我们在半监督基准数据集SVHN和CIFAR-10上证明,与MT相比,MT和性能在半义务学习中与最先进的状态相比,将我们的LC损失添加到MT中得到了显着改善。

The Mean Teacher (MT) model of Tarvainen and Valpola has shown favorable performance on several semi-supervised benchmark datasets. MT maintains a teacher model's weights as the exponential moving average of a student model's weights and minimizes the divergence between their probability predictions under diverse perturbations of the inputs. However, MT is known to suffer from confirmation bias, that is, reinforcing incorrect teacher model predictions. In this work, we propose a simple yet effective method called Local Clustering (LC) to mitigate the effect of confirmation bias. In MT, each data point is considered independent of other points during training; however, data points are likely to be close to each other in feature space if they share similar features. Motivated by this, we cluster data points locally by minimizing the pairwise distance between neighboring data points in feature space. Combined with a standard classification cross-entropy objective on labeled data points, the misclassified unlabeled data points are pulled towards high-density regions of their correct class with the help of their neighbors, thus improving model performance. We demonstrate on semi-supervised benchmark datasets SVHN and CIFAR-10 that adding our LC loss to MT yields significant improvements compared to MT and performance comparable to the state of the art in semi-supervised learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源