Sparse neural networks have received increasing interests due to their small size compared to dense networks. Nevertheless, most existing works on neural network theory have focused on dense neural networks, and our understanding of sparse networks is very limited. In this paper, we study the loss landscape of one-hidden-layer sparse networks. We first consider sparse networks with linear activations. We show that sparse linear networks can have spurious strict minima, which is in sharp contrast to dense linear networks which do not even have spurious minima. Second, we show that spurious valleys can exist for wide sparse non-linear networks. This is different from wide dense networks which do not have spurious valleys under mild assumptions.
翻译:与稠密的网络相比,松散的神经网络由于规模小而引起了越来越多的兴趣。然而,大部分现有的神经网络理论工程都集中在密集的神经网络上,我们对稀疏网络的了解非常有限。在本文中,我们研究了一层稀疏网络的流失情况。我们首先考虑的是带有线性激活的稀疏网络。我们发现,稀疏的线性网络可能具有虚假的严格微量网络,这与甚至没有虚假微量网络的稠密线性网络形成鲜明的对比。第二,我们表明,广而稀疏的非线性网络可以存在虚假的山谷。这不同于在温和假设下没有虚假的谷地的广密网络。