论文标题
通过无监督的聚类来提高半监督学习的表现
Boosting the Performance of Semi-Supervised Learning with Unsupervised Clustering
论文作者
论文摘要
最近,半监督的学习(SSL)在利用未标记的数据时表现出了很大的希望,同时提供了很少的标签。在本文中,我们表明,在训练期间,间歇性地间歇性地忽略标签可以显着提高小样本制度的性能。更具体地说,我们建议共同培训一个网络。主要的分类任务均暴露于未标记的和几乎没有注释的数据中,而次要任务则试图在没有任何标签的情况下将数据聚集。与经常用于自学的手工制作的借口任务相反,我们的聚类阶段利用了相同的分类网络并领导着放宽主要任务并在不超越标签的情况下传播来自标签的信息。最重要的是,在无监督的学习阶段进行了分类的图像旋转的自我监督技术,以稳定训练。我们证明了我们的方法在提高几种最先进的SSL算法方面的功效,显着提高了结果并减少了各种标准半监督基准的运行时间,包括CIFAR-10的92.6%精度,包括SVHN的96.9%,仅在SVHN上使用4个任务中的4个标签。我们还显着改善了每类1,2和3标签的极端情况的结果,并表明我们的模型学到的功能对于分离数据更有意义。
Recently, Semi-Supervised Learning (SSL) has shown much promise in leveraging unlabeled data while being provided with very few labels. In this paper, we show that ignoring the labels altogether for whole epochs intermittently during training can significantly improve performance in the small sample regime. More specifically, we propose to train a network on two tasks jointly. The primary classification task is exposed to both the unlabeled and the scarcely annotated data, whereas the secondary task seeks to cluster the data without any labels. As opposed to hand-crafted pretext tasks frequently used in self-supervision, our clustering phase utilizes the same classification network and head in an attempt to relax the primary task and propagate the information from the labels without overfitting them. On top of that, the self-supervised technique of classifying image rotations is incorporated during the unsupervised learning phase to stabilize training. We demonstrate our method's efficacy in boosting several state-of-the-art SSL algorithms, significantly improving their results and reducing running time in various standard semi-supervised benchmarks, including 92.6% accuracy on CIFAR-10 and 96.9% on SVHN, using only 4 labels per class in each task. We also notably improve the results in the extreme cases of 1,2 and 3 labels per class, and show that features learned by our model are more meaningful for separating the data.