论文标题
通过Distancenet-Bandits进行文本分类的多源域改编
Multi-Source Domain Adaptation for Text Classification via DistanceNet-Bandits
论文作者
论文摘要
在目标域上学习算法的域自适应性能是其源域误差的函数和这两个域的数据分布之间的差异度量。我们在NLP任务的背景下介绍了各种基于距离的度量的研究,该研究表征了基于样本估计的域之间的差异。我们首先进行分析实验,以表明这些距离测量中的哪些可以最好地将样品与同一域与不同领域区分开,并且与经验结果相关。接下来,我们开发了使用这些距离度量或这些距离度量的混合物作为额外的损耗函数,可以与任务的损失函数共同最小化,以实现更好的无监督域适应。最后,我们将该模型扩展到一个新型的伸展手式模型,该模型采用多臂Bandit控制器在多个源域之间动态切换,并允许该模型学习最佳轨迹和域的混合物,以转移到低资源目标域。我们对具有多个不同领域的流行情感分析数据集进行实验,并表明我们的Distancenet模型及其动态匪徒变体可以在无监督的域适应中胜过竞争性基线。
Domain adaptation performance of a learning algorithm on a target domain is a function of its source domain error and a divergence measure between the data distribution of these two domains. We present a study of various distance-based measures in the context of NLP tasks, that characterize the dissimilarity between domains based on sample estimates. We first conduct analysis experiments to show which of these distance measures can best differentiate samples from same versus different domains, and are correlated with empirical results. Next, we develop a DistanceNet model which uses these distance measures, or a mixture of these distance measures, as an additional loss function to be minimized jointly with the task's loss function, so as to achieve better unsupervised domain adaptation. Finally, we extend this model to a novel DistanceNet-Bandit model, which employs a multi-armed bandit controller to dynamically switch between multiple source domains and allow the model to learn an optimal trajectory and mixture of domains for transfer to the low-resource target domain. We conduct experiments on popular sentiment analysis datasets with several diverse domains and show that our DistanceNet model, as well as its dynamic bandit variant, can outperform competitive baselines in the context of unsupervised domain adaptation.