A large class of non-smooth practical optimization problems can be written as minimization of a sum of smooth and partly smooth functions. We consider such structured problems which also depend on a parameter vector and study the problem of differentiating its solution mapping with respect to the parameter which has far reaching applications in sensitivity analysis and parameter learning optmization problems. We show that under partial smoothness and other mild assumptions, Automatic Differentiation (AD) of the sequence generated by proximal splitting algorithms converges to the derivative of the solution mapping. For a variant of automatic differentiation, which we call Fixed-Point Automatic Differentiation (FPAD), we remedy the memory overhead problem of the Reverse Mode AD and moreover provide faster convergence theoretically. We numerically illustrate the convergence and convergence rates of AD and FPAD on Lasso and Group Lasso problems and demonstrate the working of FPAD on prototypical practical image denoising problem by learning the regularization term.
翻译:大量非悬浮实际优化问题可以写成最大限度地减少平滑和部分平滑功能的总和。我们考虑了这些结构化问题,这些问题也取决于参数矢量,研究如何将其解决方案绘图与在敏感度分析和参数学习选择化问题方面影响深远的参数区分开来的问题。我们表明,在部分平滑和其他温和假设下,近似分解算法产生的序列自动差异(AD)与解决方案映射的衍生物汇合在一起。关于自动区分的变体,我们称之为固定点自动区分(FPAD),我们纠正反向模式的记忆间接问题,并在理论上提供更快的趋同。我们用数字说明反倾销和FPAD在Lasso和Lasso组问题上的趋同率和趋同率,并通过学习正规化术语,展示了FPAD在原型实际图像脱色问题方面的工作。