论文标题
受环球启发的监督对比学习
Universum-inspired Supervised Contrastive Learning
论文作者
论文摘要
作为一种有效的数据增强方法,混合通过线性插值合成了额外的样品。尽管理论上依赖数据属性,但据报道,Mixup的表现良好,并且是正规器和校准器,可以促进可靠的鲁棒性和对深度模型训练的概括。在本文中,我们受到使用类样本来协助目标任务的Universum学习的启发,我们从很大程度上探索的视角研究了混音,这是生成不属于目标类别的域内样本的潜力。我们发现,在有监督的对比学习的框架内,混合引起的大学可以充当令人惊讶的高质量艰苦的负面负面因素,从而极大地减轻了对比偏学习中对大批量大小的需求。通过这些发现,我们提出了受青睐的受监督的对比学习(UNICON),该学习结合了混合策略,以产生混合诱导的大学作为农业,并将其与目标类别的锚定样品分开。我们将我们的方法扩展到无监督的设置,提出了无监督的对比模型(UN-UNI)。我们的方法不仅可以改善与硬标签的混合,还可以创新一种新的措施来生成Universum数据。 Unicon凭借有关学习表示的线性分类器,在各种数据集上显示了最先进的性能。特别是,Unicon在CIFAR-100上获得了81.7%的TOP-1准确性,超过了5.2%的明显利润率,批量较小,通常使用Resnet-50在UNICON和SUPCON中的1024中,通常为256。 UN-UNI在CIFAR-100上也胜过SOTA方法。本文的代码在https://github.com/hannaiiyanggit/unicon上发布。
As an effective data augmentation method, Mixup synthesizes an extra amount of samples through linear interpolations. Despite its theoretical dependency on data properties, Mixup reportedly performs well as a regularizer and calibrator contributing reliable robustness and generalization to deep model training. In this paper, inspired by Universum Learning which uses out-of-class samples to assist the target tasks, we investigate Mixup from a largely under-explored perspective - the potential to generate in-domain samples that belong to none of the target classes, that is, universum. We find that in the framework of supervised contrastive learning, Mixup-induced universum can serve as surprisingly high-quality hard negatives, greatly relieving the need for large batch sizes in contrastive learning. With these findings, we propose Universum-inspired supervised Contrastive learning (UniCon), which incorporates Mixup strategy to generate Mixup-induced universum as universum negatives and pushes them apart from anchor samples of the target classes. We extend our method to the unsupervised setting, proposing Unsupervised Universum-inspired contrastive model (Un-Uni). Our approach not only improves Mixup with hard labels, but also innovates a novel measure to generate universum data. With a linear classifier on the learned representations, UniCon shows state-of-the-art performance on various datasets. Specially, UniCon achieves 81.7% top-1 accuracy on CIFAR-100, surpassing the state of art by a significant margin of 5.2% with a much smaller batch size, typically, 256 in UniCon vs. 1024 in SupCon using ResNet-50. Un-Uni also outperforms SOTA methods on CIFAR-100. The code of this paper is released on https://github.com/hannaiiyanggit/UniCon.