Implicit-depth models such as Deep Equilibrium Networks have recently been shown to match or exceed the performance of traditional deep networks while being much more memory efficient. However, these models suffer from unstable convergence to a solution and lack guarantees that a solution exists. On the other hand, Neural ODEs, another class of implicit-depth models, do guarantee existence of a unique solution but perform poorly compared with traditional networks. In this paper, we develop a new class of implicit-depth model based on the theory of monotone operators, the Monotone Operator Equilibrium Network (monDEQ). We show the close connection between finding the equilibrium point of an implicit network and solving a form of monotone operator splitting problem, which admits efficient solvers with guaranteed, stable convergence. We then develop a parameterization of the network which ensures that all operators remain monotone, which guarantees the existence of a unique equilibrium point. Finally, we show how to instantiate several versions of these models, and implement the resulting iterative solvers, for structured linear operators such as multi-scale convolutions. The resulting models vastly outperform the Neural ODE-based models while also being more computationally efficient. Code is available at http://github.com/locuslab/monotone_op_net.
翻译:深平衡网络等隐含深度模型最近被证明与传统深网络的性能相匹配或超过,而其记忆效率则更高。然而,这些模型存在不稳定的趋同,缺乏解决方案存在保障。另一方面,另一类隐含深度模型神经内分解器(Neural ODEs)确实保障存在一个独特的解决方案,但与传统网络相比却表现不佳。在本文件中,我们根据单体操作器理论,即单体操作器电子平衡网络(monDEQ),开发了一种新的隐含深度模型。我们展示了在找到一个隐性网络的平衡点与解决单一操作器分裂问题之间,以及解决一种单一操作器分解问题的形式之间的密切联系,后者承认了高效的求同器,且稳定地趋同。然后我们开发了网络的参数化,以确保所有操作器都保持单一,从而保证存在一个独特的平衡点。最后,我们展示了如何将这些模型的若干版本瞬间化,并为结构化的线性操作器,例如多尺度的演算器。由此产生的模型大大超越了可保证、稳定的内核/内存式模型。同时,在更高效的内存式的计算中。