论文标题
HMOE:用于领域概括的专家的高网络混合物
HMOE: Hypernetwork-based Mixture of Experts for Domain Generalization
论文作者
论文摘要
由于领域的变化,机器学习系统通常很难将与培训数据不同的新领域充分推广,这是域泛化(DG)的目的。尽管已经提出了各种DG方法,但其中大多数的可解释性均缺乏,并且需要域标签,这在许多现实世界中都不可用。本文提出了一种新型的DG方法,称为HMOE:基于超网络的专家混合物(MOE),它不依赖域标签,并且更容易解释。 MOE证明有效地识别数据中的异质模式。对于DG问题,异质性完全来自域移位。 HMOE采用超级核武器将向量作为输入来产生专家的权重,从而促进了专家之间的知识共享,并可以在低维矢量空间中探索其相似之处。在公平评估框架下,我们基准针对其他DG方法 - 域名。我们的广泛实验表明,HMOE可以有效地将混合域数据分离为不同的簇,这些簇与人类直觉比原始域标签更一致。 HMOE使用自我学习的域信息,在大多数数据集上实现最先进的结果,并且在所有数据集中平均超过了其他DG方法。
Due to domain shifts, machine learning systems typically struggle to generalize well to new domains that differ from those of training data, which is what domain generalization (DG) aims to address. Although a variety of DG methods have been proposed, most of them fall short in interpretability and require domain labels, which are not available in many real-world scenarios. This paper presents a novel DG method, called HMOE: Hypernetwork-based Mixture of Experts (MoE), which does not rely on domain labels and is more interpretable. MoE proves effective in identifying heterogeneous patterns in data. For the DG problem, heterogeneity arises exactly from domain shifts. HMOE employs hypernetworks taking vectors as input to generate the weights of experts, which promotes knowledge sharing among experts and enables the exploration of their similarities in a low-dimensional vector space. We benchmark HMOE against other DG methods under a fair evaluation framework -- DomainBed. Our extensive experiments show that HMOE can effectively separate mixed-domain data into distinct clusters that are surprisingly more consistent with human intuition than original domain labels. Using self-learned domain information, HMOE achieves state-of-the-art results on most datasets and significantly surpasses other DG methods in average accuracy across all datasets.