论文标题
关于在神经网络中使用Word2VEC表示对话法的影响
On the Effects of Using word2vec Representations in Neural Networks for Dialogue Act Recognition
论文作者
论文摘要
对话法案识别是大量自然语言处理管道的重要组成部分。在这一领域进行了许多研究工作,但相对较少的研究深度神经网络和单词嵌入。鉴于这两种技术在大多数其他与语言有关的领域中都非常好,这是令人惊讶的。我们在这项工作中提出了一个新的深神经网络,该网络探讨了经常性模型以捕获句子中的单词序列,并进一步研究了验证的单词嵌入的影响。我们用三种语言来验证这种模型:英语,法语和捷克语。这些语言之间提出的方法的性能是一致的,并且与英语的最新结果相媲美。更重要的是,我们确认深神经网络确实超过了最大熵分类器,这是可以预期的。但是,这更令人惊讶,我们还发现,无论培训语料库的大小如何,标准Word2Vec EM床单似乎都不会为该任务和拟议模型带来宝贵的信息。因此,我们进一步分析了所得的嵌入,并得出结论,可能的解释可能与Word2Vec嵌入术中捕获的词汇语义信息的类型之间的不匹配有关,以及对对话对话ACT识别任务最有用的单词之间的关系类型。
Dialogue act recognition is an important component of a large number of natural language processing pipelines. Many research works have been carried out in this area, but relatively few investigate deep neural networks and word embeddings. This is surprising, given that both of these techniques have proven exceptionally good in most other language-related domains. We propose in this work a new deep neural network that explores recurrent models to capture word sequences within sentences, and further study the impact of pretrained word embeddings. We validate this model on three languages: English, French and Czech. The performance of the proposed approach is consistent across these languages and it is comparable to the state-of-the-art results in English. More importantly, we confirm that deep neural networks indeed outperform a Maximum Entropy classifier, which was expected. However , and this is more surprising, we also found that standard word2vec em-beddings do not seem to bring valuable information for this task and the proposed model, whatever the size of the training corpus is. We thus further analyse the resulting embeddings and conclude that a possible explanation may be related to the mismatch between the type of lexical-semantic information captured by the word2vec embeddings, and the kind of relations between words that is the most useful for the dialogue act recognition task.