Implicit equilibrium models, i.e., deep neural networks (DNNs) defined by implicit equations, have been becoming more and more attractive recently. In this paper, we investigate an emerging question: can an implicit equilibrium model's equilibrium point be regarded as the solution of an optimization problem? To this end, we first decompose DNNs into a new class of unit layer that is the proximal operator of an implicit convex function while keeping its output unchanged. Then, the equilibrium model of the unit layer can be derived, named Optimization Induced Equilibrium Networks (OptEq), which can be easily extended to deep layers. The equilibrium point of OptEq can be theoretically connected to the solution of its corresponding convex optimization problem with explicit objectives. Based on this, we can flexibly introduce prior properties to the equilibrium points: 1) modifying the underlying convex problems explicitly so as to change the architectures of OptEq; and 2) merging the information into the fixed point iteration, which guarantees to choose the desired equilibrium point when the fixed point set is non-singleton. We show that deep OptEq outperforms previous implicit models even with fewer parameters. This work establishes the first step towards the optimization-guided design of deep models.
翻译:隐含的平衡模型,即由隐含方程式定义的深神经网络(DNN),最近越来越具有吸引力。在本文件中,我们调查了一个新出现的问题:隐含的平衡模型的平衡点能否被视为优化问题的解决方案?为此目的,我们首先将DNN转换成一个新的单元层类别,即隐含二次曲线功能的精密操作器,同时保持其输出不变。然后,可以得出单位层的平衡模型,名为优化引导平衡网络(OptEq),它很容易扩展至深层。OptEq的平衡点可以在理论上与其相应的矩形优化问题的解决方案相连接,并具有明确的目标。在此基础上,我们可以灵活地将先前的属性引入平衡点:1) 修改基本的矩轴问题,以明确改变 OptEq 的结构;2) 将信息合并到固定的循环点,这将保证在固定点设定的深度平衡点上选择理想的平衡点(OptEq),即使固定的模型是非隐含的优化模型,也保证选择了前一个更深的模型。