Neural networks (NNs) struggle to efficiently solve certain problems, such as learning parities, even when there are simple learning algorithms for those problems. Can NNs discover learning algorithms on their own? We exhibit a NN architecture that, in polynomial time, learns as well as any efficient learning algorithm describable by a constant-sized program. For example, on parity problems, the NN learns as well as Gaussian elimination, an efficient algorithm that can be succinctly described. Our architecture combines both recurrent weight sharing between layers and convolutional weight sharing to reduce the number of parameters down to a constant, even though the network itself may have trillions of nodes. While in practice the constants in our analysis are too large to be directly meaningful, our work suggests that the synergy of Recurrent and Convolutional NNs (RCNNs) may be more natural and powerful than either alone, particularly for concisely parameterizing discrete algorithms.
翻译:神经网络(NNs)努力有效地解决某些问题,例如学习平等,即使这些问题有简单的学习算法。NNS能够自己发现学习算法吗?我们展示了NNS结构,在多元时间里,学习以及任何可以被一个不变规模的程序所剥夺的有效学习算法。例如,关于均等问题,NN学会了高萨的消灭,这是一种可以简单描述的高效算法。我们的架构将各层之间经常的权重分担和革命权重分享结合起来,将参数数量降低到恒定水平,即使网络本身可能有数万亿个节点。在实践中,我们分析中的常数太大,无法直接产生意义,但我们的工作表明,经常性和革命性NNS(RCNN)的协同作用可能比单独更自然、更强大,特别是简洁的参数化离散算法。