Out of the recent advances in systems and control (S\&C)-based analysis of optimization algorithms, not enough work has been specifically dedicated to machine learning (ML) algorithms and its applications. This paper addresses this gap by illustrating how (discrete-time) Lyapunov stability theory can serve as a powerful tool to aid, or even lead, in the analysis (and potential design) of optimization algorithms that are not necessarily gradient-based. The particular ML problem that this paper focuses on is that of parameter estimation in an incomplete-data Bayesian framework via the popular optimization algorithm known as maximum a posteriori expectation-maximization (MAP-EM). Following first principles from dynamical systems stability theory, conditions for convergence of MAP-EM are developed. Furthermore, if additional assumptions are met, we show that fast convergence (linear or quadratic) is achieved, which could have been difficult to unveil without our adopted S\&C approach. The convergence guarantees in this paper effectively expand the set of sufficient conditions for EM applications, thereby demonstrating the potential of similar S\&C-based convergence analysis of other ML algorithms.
翻译:在基于系统和控制的优化算法分析(S ⁇ C)的最新进展中,没有专门为机器学习(ML)算法及其应用专门开展足够的工作,本文件通过说明Lyapunov稳定性理论如何(分流-时间)在分析(和潜在设计)不一定基于梯度的优化算法分析(不一定基于梯度的优化算法分析(和潜在设计)中起到帮助或甚至引导作用,从而说明Lyapunov稳定性理论如何(分流-时间)作为一个强有力的工具来帮助甚至引导这一差距。本文着重讨论的特定的ML问题是,通过被称为后期预期-最大程度的普及优化算法(MAP-EM),在一个不完整的Bayesian框架内对参数进行估计,从而展示了以S ⁇ C为基础的其他ML算法趋同分析的可能性。此外,如果满足了其他假设,我们表明,实现快速趋同(线性或四面式)是难以实现的。