A recent work by Ramanujan et al. (2020) provides significant empirical evidence that sufficiently overparameterized, random neural networks contain untrained subnetworks that achieve state-of-the-art accuracy on several predictive tasks. A follow-up line of theoretical work provides justification of these findings by proving that slightly overparameterized neural networks, with commonly used continuous-valued random initializations can indeed be pruned to approximate any target network. In this work, we show that the amplitude of those random weights does not even matter. We prove that any target network can be approximated up to arbitrary accuracy by simply pruning a random network of binary $\{\pm1\}$ weights that is only a polylogarithmic factor wider and deeper than the target network.
翻译:Ramanujan等人(2020年)最近的一项工作提供了重要的实证证据,证明足够超度的随机神经网络包含未经训练的亚网络,这些子网络在几项预测任务上达到最先进的精确度。 一项后续的理论工作为这些结论提供了理由,证明略微超度的超度神经网络,以及常用的连续估值随机初始化,确实可以被切割到接近任何目标网络。 在这项工作中,我们证明这些随机重量的振幅甚至无关紧要。 我们证明,任何目标网络都可以通过简单剪裁一个仅比目标网络更广泛和更深的多logalariphical $ ⁇ pm1 ⁇ $$$$$$$$$$$的随机网络来接近任意精确度。