This paper revisits Deep Mutual Learning (DML), a simple yet effective computing paradigm. We propose using R\'{e}nyi divergence instead of the KL divergence, which is more flexible and tunable, to improve vanilla DML. This modification is able to consistently improve performance over vanilla DML with limited additional complexity. The convergence properties of the proposed paradigm are analyzed theoretically, and Stochastic Gradient Descent with a constant learning rate is shown to converge with $\mathcal{O}(1)$-bias in the worst case scenario for nonconvex optimization tasks. That is, learning will reach nearby local optima but continue searching within a bounded scope, which may help mitigate overfitting. Finally, our extensive empirical results demonstrate the advantage of combining DML and R\'{e}nyi divergence, which further improves generalized models.
翻译:本文重新审视了深度互学习(DML)这一简单而有效的计算范型。我们提出使用 R\'{e}nyi 散度代替 KL 散度,后者更加灵活和可调节,以改进基本 DML。这种修改能够通过有限的额外复杂性持续提高性能。对所提出范型的收敛性进行了理论分析,证明了具有常数学习率的随机梯度下降法在非凸优化任务的最坏情况下会收敛到 $\mathcal{O}(1)$ 偏差。也就是,学习将到达附近的局部最优解,但在有限范围内继续搜索,这可能有助于缓解过拟合。最后,我们广泛的实证结果证明了结合 DML 和 R\'{e}nyi 散度的优势,使广义模型进一步提高。