This paper concludes a series of studies on the polyharmonic cascade, a deep machine learning architecture theoretically derived from indifference principles and the theory of random functions. A universal initialization procedure is proposed, based on symmetric constellations in the form of hyperoctahedra with a central point. This initialization not only ensures stable training of cascades with tens and hundreds of layers (up to 500 layers without skip connections), but also radically simplifies the computations. Scalability and robustness are demonstrated on MNIST (98.3% without convolutions or augmentations), HIGGS (AUC approximately 0.885 on 11M examples), and Epsilon (AUC approximately 0.963 with 2000 features). All linear algebra is reduced to 2D operations and is efficiently executed on GPUs. A public repository and an archived snapshot are provided for full reproducibility.


翻译:本文是关于多谐级联(一种基于无差别原理和随机函数理论推导出的深度学习架构)系列研究的总结。提出了一种基于带中心点的超八面体对称构型的通用初始化方法。该初始化不仅能确保数十乃至数百层(无跳跃连接下可达500层)级联的稳定训练,还能从根本上简化计算。在MNIST(无卷积或数据增强下准确率达98.3%)、HIGGS(1100万样本上AUC约0.885)和Epsilon(2000维特征下AUC约0.963)数据集上验证了其可扩展性与鲁棒性。所有线性代数运算均简化为二维操作,并可在GPU上高效执行。为保障完全可复现性,已提供公开代码库及归档快照。

0
下载
关闭预览

相关内容

Top
微信扫码咨询专知VIP会员