论文标题
神经跨语性摘要的变分等级模型
A Variational Hierarchical Model for Neural Cross-Lingual Summarization
论文作者
论文摘要
跨语性摘要(CLS)的目标是将一种语言(例如英语)转换为另一种语言(例如中文)中的摘要。本质上,CLS任务是机器翻译(MT)和单语摘要(MS)的组合,因此存在MT \&MS和CLS之间的层次关系。现有对CLS的研究主要集中于利用管道方法或通过辅助MT或MS目标共同训练端到端模型。但是,模型直接进行CLS非常具有挑战性,因为它需要翻译和总结的能力。为了解决此问题,我们根据条件变异自动编码器为CLS任务提出了一个层次模型。分层模型分别在本地和全局级别包含两种潜在变量。在本地层面,有两个潜在变量,一个用于翻译,另一个用于摘要。至于全球级别,在两个局部级变量上有一个跨语义汇总的潜在变量。在两个语言方向(英语 - 中国)上进行实验验证了所提出的方法的有效性和优势。此外,我们表明我们的模型能够产生比在几个拍摄设置中的比较模型更好的跨语性摘要。
The goal of the cross-lingual summarization (CLS) is to convert a document in one language (e.g., English) to a summary in another one (e.g., Chinese). Essentially, the CLS task is the combination of machine translation (MT) and monolingual summarization (MS), and thus there exists the hierarchical relationship between MT\&MS and CLS. Existing studies on CLS mainly focus on utilizing pipeline methods or jointly training an end-to-end model through an auxiliary MT or MS objective. However, it is very challenging for the model to directly conduct CLS as it requires both the abilities to translate and summarize. To address this issue, we propose a hierarchical model for the CLS task, based on the conditional variational auto-encoder. The hierarchical model contains two kinds of latent variables at the local and global levels, respectively. At the local level, there are two latent variables, one for translation and the other for summarization. As for the global level, there is another latent variable for cross-lingual summarization conditioned on the two local-level variables. Experiments on two language directions (English-Chinese) verify the effectiveness and superiority of the proposed approach. In addition, we show that our model is able to generate better cross-lingual summaries than comparison models in the few-shot setting.