Structure learning of directed acyclic graphs (DAGs) is a fundamental problem in many scientific endeavors. A new line of work, based on NOTEARS (Zheng et al., 2018), reformulates the structure learning problem as a continuous optimization one by leveraging an algebraic characterization of DAG constraint. The constrained problem is typically solved using the augmented Lagrangian method (ALM) which is often preferred to the quadratic penalty method (QPM) by virtue of its convergence result that does not require the penalty coefficient to go to infinity, hence avoiding ill-conditioning. In this work, we review the standard convergence result of the ALM and show that the required conditions are not satisfied in the recent continuous constrained formulation for learning DAGs. We demonstrate empirically that its behavior is akin to that of the QPM which is prone to ill-conditioning, thus motivating the use of second-order method in this setting. We also establish the convergence guarantee of QPM to a DAG solution, under mild conditions, based on a property of the DAG constraint term.
翻译:定向自行车图(DAGs)的结构学习是许多科学工作的一个根本问题。根据NOTARRS(Zheng等人,2018年),通过利用DAG限制的代数特征,将结构学习问题重新定位为连续优化问题,通过利用对DAG限制的代数特征,将结构学习问题重新定位为连续优化问题。受限制问题通常通过扩大Lagrangian方法(ALM)来解决,该方法往往偏向于四级惩罚方法(QPM),因为其趋同结果并不要求惩罚系数达到无限程度,从而避免了不良的调节。在这项工作中,我们审查了ALM的标准趋同结果,并表明在最近为学习DAG的受限制的配方中,没有满足所需的条件。我们从经验上表明,它的行为类似于QPM(QPM),它容易受到不当调节,从而在这种环境下鼓励使用二级惩罚方法。我们还根据DAG限制条件的特性,在较轻的条件下,建立了QPM的趋同性保证。