In light of the smoothness property brought by skip connections in ResNet, this paper proposed the Skip Logit to introduce the skip connection mechanism that fits arbitrary DNN dimensions and embraces similar properties to ResNet. Meta Tanh Normalization (MTN) is designed to learn variance information and stabilize the training process. With these delicate designs, our Skip Meta Logit (SML) brought incremental boosts to the performance of extensive SOTA ctr prediction models on two real-world datasets. In the meantime, we prove that the optimization landscape of arbitrarily deep skip logit networks has no spurious local optima. Finally, SML can be easily added to building blocks and has delivered offline accuracy and online business metrics gains on app ads learning to rank systems at TikTok.
翻译:鉴于ResNet中跳过连接带来的平滑性能,本文建议跳过登录系统引入符合任意 DNN 维度并包含ResNet类似特性的跳过连接机制。 Meta Tanh 正常化(MTN)旨在学习差异信息并稳定培训过程。有了这些微妙的设计,我们的跳过Meta Logit(SML)在两个真实世界数据集上为广泛的SOTA ctr预测模型的运行带来了增量。 与此同时,我们证明任意深度跳过登录网络的优化景观没有虚假的本地opima。 最后, SML可以很容易地添加到构建区块中,并在TikTok的应用程序上提供离线的准确性和在线商业指标。