论文标题
Mose:多模式知识图完成的模态拆分和合奏
MoSE: Modality Split and Ensemble for Multimodal Knowledge Graph Completion
论文作者
论文摘要
多模式知识图完成(MKGC)旨在预测MKG中缺失的实体。以前的工作通常跨模式共享关系表示。这会导致训练过程中的方式之间的相互干扰,因为对于一对实体,一种模式的关系可能与另一种方式相矛盾。此外,根据共享关系表示做出统一的预测,以不同的方式对待输入,而其对MKGC任务的重要性应不同。在本文中,我们提出了Mose的建议,即MKGC的模态拆分表示和集合推理框架。具体而言,在训练阶段,我们学习了每种模态的模态分裂关系嵌入,而不是单个模态共享的嵌入,从而减轻了模态干扰。基于这些嵌入,在推理阶段,我们首先进行模态分解预测,然后利用各种集合方法将预测与不同权重相结合,这将动态地模拟模态重要性。三个公斤数据集的实验结果表明,摩西的表现优于最先进的MKGC方法。代码可在https://github.com/oreozhao/mose4mkgc上找到。
Multimodal knowledge graph completion (MKGC) aims to predict missing entities in MKGs. Previous works usually share relation representation across modalities. This results in mutual interference between modalities during training, since for a pair of entities, the relation from one modality probably contradicts that from another modality. Furthermore, making a unified prediction based on the shared relation representation treats the input in different modalities equally, while their importance to the MKGC task should be different. In this paper, we propose MoSE, a Modality Split representation learning and Ensemble inference framework for MKGC. Specifically, in the training phase, we learn modality-split relation embeddings for each modality instead of a single modality-shared one, which alleviates the modality interference. Based on these embeddings, in the inference phase, we first make modality-split predictions and then exploit various ensemble methods to combine the predictions with different weights, which models the modality importance dynamically. Experimental results on three KG datasets show that MoSE outperforms state-of-the-art MKGC methods. Codes are available at https://github.com/OreOZhao/MoSE4MKGC.