论文标题
正确的归一化事项:了解归一化对点击率预测深度神经网络模型的影响
Correct Normalization Matters: Understanding the Effect of Normalization On Deep Neural Network Models For Click-Through Rate Prediction
论文作者
论文摘要
正常化已成为许多深层神经网络中用于机器学习任务的最基本组件之一,而深层神经网络也已被广泛用于CTR估计领域。在大多数提出的深神经网络模型中,很少有模型使用归一化方法。尽管某些作品(例如深层和跨网络(DCN)和神经分解机器(NFM)在结构的MLP一部分中都使用批准化,但没有工作可以彻底探索归一化对DNN排名系统的影响。在本文中,我们通过将各种归一化方法应用于DNN模型中的特征嵌入和MLP部分,对广泛使用归一化模式的影响进行系统研究。在三个现实世界数据集上进行了广泛的实验,实验结果表明,正确的归一化显着增强了模型的性能。我们还提出了一种基于在这项工作中仅名为“差异”的分层方差(vo-ln)的新的有效归一化方法。还基于上述观察结果提出了标准化增强的DNN模型。至于归一化在CTR估计中适用于DNN模型的原因,我们发现归一化的差异起着主要作用,并在这项工作中给出了解释。
Normalization has become one of the most fundamental components in many deep neural networks for machine learning tasks while deep neural network has also been widely used in CTR estimation field. Among most of the proposed deep neural network models, few model utilize normalization approaches. Though some works such as Deep & Cross Network (DCN) and Neural Factorization Machine (NFM) use Batch Normalization in MLP part of the structure, there isn't work to thoroughly explore the effect of the normalization on the DNN ranking systems. In this paper, we conduct a systematic study on the effect of widely used normalization schemas by applying the various normalization approaches to both feature embedding and MLP part in DNN model. Extensive experiments are conduct on three real-world datasets and the experiment results demonstrate that the correct normalization significantly enhances model's performance. We also propose a new and effective normalization approaches based on LayerNorm named variance only LayerNorm(VO-LN) in this work. A normalization enhanced DNN model named NormDNN is also proposed based on the above-mentioned observation. As for the reason why normalization works for DNN models in CTR estimation, we find that the variance of normalization plays the main role and give an explanation in this work.