A recently proposed method in deep learning groups multiple neurons to capsules such that each capsule represents an object or part of an object. Routing algorithms route the output of capsules from lower-level layers to upper-level layers. In this paper, we prove that state-of-the-art routing procedures decrease the expressivity of capsule networks. More precisely, it is shown that EM-routing and routing-by-agreement prevent capsule networks from distinguishing inputs and their negative counterpart. Therefore, only symmetric functions can be expressed by capsule networks, and it can be concluded that they are not universal approximators. We also theoretically motivate and empirically show that this limitation affects the training of deep capsule networks negatively. Therefore, we present an incremental improvement for state-of-the-art routing algorithms that solves the aforementioned limitation and stabilizes the training of capsule networks.
翻译:在深层学习组群中,最近提出的一种方法是多个神经元到胶囊,使每个胶囊代表一个物体或一个物体的一部分。运行算法将胶囊的输出从低层到高层。在本文中,我们证明最先进的路由程序降低了胶囊网络的表达性。更确切地说,EM路由和路由逐个协议使得胶囊网络无法区分投入及其负对等功能。因此,只有对称功能才能通过胶囊网络来表达,并且可以得出结论,它们不是通用的近似器。我们从理论上激励和从经验上表明,这种限制对深层胶囊网络的培训产生了负面影响。因此,我们对解决上述限制和稳定胶囊网络培训的最先进路由算法提出了渐进的改进。