论文标题
负责通过人类的同伴学习
Responsible Active Learning via Human-in-the-loop Peer Study
论文作者
论文摘要
已经提出了主动学习来减少数据注释工作,仅通过手动将代表性数据样本标记进行培训。同时,最近的活跃学习应用程序从具有足够的计算资源的云计算服务中受益匪浅,而且还受益于众包框架,这些框架在主动学习循环中包括许多人。但是,以前的活跃学习方法总是需要将大规模的未标记数据传递给云,可能会引发重大数据隐私问题。为了减轻这种风险,我们提出了一种负责任的主动学习方法,即同行学习学习(PSL),以同时保留数据隐私并提高模型稳定性。具体来说,我们首先引入了一个人类的教师学生建筑,以通过在客户端维护活跃的学习者(学生),从云方面隔离任务学习者(教师)的未标记数据。在培训期间,任务学习者指示轻量级活跃的学习者,然后提供有关主动采样标准的反馈。为了通过大规模的未标记数据进一步增强活跃的学习者,我们将多个同伴学生介绍给活跃的学习者,该学习者受到新的学习范式培训,包括对标签数据的课堂同行研究和对未标记数据的课外同伴研究。最后,我们设计了一个基于差异的主动抽样标准,同行研究反馈,该标准利用同伴学生选择最有用的数据以提高模型稳定性。广泛的实验表明,在标准和敏感保护设置中,所提出的PSL比广泛的主动学习方法具有优势。
Active learning has been proposed to reduce data annotation efforts by only manually labelling representative data samples for training. Meanwhile, recent active learning applications have benefited a lot from cloud computing services with not only sufficient computational resources but also crowdsourcing frameworks that include many humans in the active learning loop. However, previous active learning methods that always require passing large-scale unlabelled data to cloud may potentially raise significant data privacy issues. To mitigate such a risk, we propose a responsible active learning method, namely Peer Study Learning (PSL), to simultaneously preserve data privacy and improve model stability. Specifically, we first introduce a human-in-the-loop teacher-student architecture to isolate unlabelled data from the task learner (teacher) on the cloud-side by maintaining an active learner (student) on the client-side. During training, the task learner instructs the light-weight active learner which then provides feedback on the active sampling criterion. To further enhance the active learner via large-scale unlabelled data, we introduce multiple peer students into the active learner which is trained by a novel learning paradigm, including the In-Class Peer Study on labelled data and the Out-of-Class Peer Study on unlabelled data. Lastly, we devise a discrepancy-based active sampling criterion, Peer Study Feedback, that exploits the variability of peer students to select the most informative data to improve model stability. Extensive experiments demonstrate the superiority of the proposed PSL over a wide range of active learning methods in both standard and sensitive protection settings.