论文标题

关于结构学习的连续约束优化的收敛

On the Convergence of Continuous Constrained Optimization for Structure Learning

论文作者

Ng, Ignavier, Lachapelle, Sébastien, Ke, Nan Rosemary, Lacoste-Julien, Simon, Zhang, Kun

论文摘要

最近,通过利用代数表征无环的结构学习,已将定向无环图(DAG)作为连续优化问题提出。使用增强的拉格朗日方法(ALM)解决了约束问题,该方法通常优先于二次罚款方法(qpm),这是由于其标准收敛结果,而该结果不需要罚款系数即可到达无限,因此避免了不良条件。但是,这些用于结构学习的方法的收敛属性,包括保证是否可以返回DAG解决方案,尚不清楚,这可能会限制其实际应用。在这项工作中,我们检查了ALM和QPM在线性,非线性和混淆情况下进行结构学习的收敛性。我们表明,在这些环境中,ALM的标准收敛结果并不成立,并从经验上证明其行为类似于QPM的行为,QPM容易发生不良。在轻度条件下,我们进一步建立了QPM与DAG溶液的收敛保证。最后,我们将理论结果与现有方法联系起来,以帮助解决融合问题,并根据对它们的经验比较来验证我们的发现。

Recently, structure learning of directed acyclic graphs (DAGs) has been formulated as a continuous optimization problem by leveraging an algebraic characterization of acyclicity. The constrained problem is solved using the augmented Lagrangian method (ALM) which is often preferred to the quadratic penalty method (QPM) by virtue of its standard convergence result that does not require the penalty coefficient to go to infinity, hence avoiding ill-conditioning. However, the convergence properties of these methods for structure learning, including whether they are guaranteed to return a DAG solution, remain unclear, which might limit their practical applications. In this work, we examine the convergence of ALM and QPM for structure learning in the linear, nonlinear, and confounded cases. We show that the standard convergence result of ALM does not hold in these settings, and demonstrate empirically that its behavior is akin to that of the QPM which is prone to ill-conditioning. We further establish the convergence guarantee of QPM to a DAG solution, under mild conditions. Lastly, we connect our theoretical results with existing approaches to help resolve the convergence issue, and verify our findings in light of an empirical comparison of them.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源